[ 544.064782] env[61986]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61986) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 544.065265] env[61986]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61986) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 544.065265] env[61986]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61986) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 544.065515] env[61986]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 544.161907] env[61986]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61986) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 544.171975] env[61986]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61986) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 544.773088] env[61986]: INFO nova.virt.driver [None req-442c9341-5ae2-4b21-9618-8480e4042607 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 544.843086] env[61986]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.846027] env[61986]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.846027] env[61986]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61986) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 547.959383] env[61986]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-19d25588-0cf4-4f68-8abb-2b1252a695f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.975061] env[61986]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61986) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 547.975208] env[61986]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-cc8752ec-b964-4d44-9e03-dbe1a17fd6fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.999448] env[61986]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 26e32. [ 547.999595] env[61986]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.156s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.000103] env[61986]: INFO nova.virt.vmwareapi.driver [None req-442c9341-5ae2-4b21-9618-8480e4042607 None None] VMware vCenter version: 7.0.3 [ 548.003502] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d33b16-a027-43b4-979f-013f013abd72 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.025290] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c086cf9-c556-47dd-b57e-0c0bbac5e5a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.031239] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe7f46b-04de-453d-b226-99838bbe5acd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.037799] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16187ed-cbb1-4d9b-ba97-f7c8ccbce305 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.050908] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2198c80b-c328-486b-a301-bd33a87cf581 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.056721] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378a8f69-43e4-47cd-bbda-a1e094473fe2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.086248] env[61986]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-97068761-f1b2-45b0-a9a6-92c6f9bc1a33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.091234] env[61986]: DEBUG nova.virt.vmwareapi.driver [None req-442c9341-5ae2-4b21-9618-8480e4042607 None None] Extension org.openstack.compute already exists. {{(pid=61986) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:224}} [ 548.093923] env[61986]: INFO nova.compute.provider_config [None req-442c9341-5ae2-4b21-9618-8480e4042607 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 548.597539] env[61986]: DEBUG nova.context [None req-442c9341-5ae2-4b21-9618-8480e4042607 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),f9811cd0-3c0d-4d0c-bd6c-7affc09d0836(cell1) {{(pid=61986) load_cells /opt/stack/nova/nova/context.py:464}} [ 548.600611] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.600841] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.601625] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.602063] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Acquiring lock "f9811cd0-3c0d-4d0c-bd6c-7affc09d0836" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.602256] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Lock "f9811cd0-3c0d-4d0c-bd6c-7affc09d0836" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.603290] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Lock "f9811cd0-3c0d-4d0c-bd6c-7affc09d0836" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.624457] env[61986]: INFO dbcounter [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Registered counter for database nova_cell0 [ 548.633011] env[61986]: INFO dbcounter [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Registered counter for database nova_cell1 [ 548.636339] env[61986]: DEBUG oslo_db.sqlalchemy.engines [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61986) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 548.637050] env[61986]: DEBUG oslo_db.sqlalchemy.engines [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61986) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 548.641812] env[61986]: ERROR nova.db.main.api [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.641812] env[61986]: result = function(*args, **kwargs) [ 548.641812] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.641812] env[61986]: return func(*args, **kwargs) [ 548.641812] env[61986]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 548.641812] env[61986]: result = fn(*args, **kwargs) [ 548.641812] env[61986]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 548.641812] env[61986]: return f(*args, **kwargs) [ 548.641812] env[61986]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 548.641812] env[61986]: return db.service_get_minimum_version(context, binaries) [ 548.641812] env[61986]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 548.641812] env[61986]: _check_db_access() [ 548.641812] env[61986]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 548.641812] env[61986]: stacktrace = ''.join(traceback.format_stack()) [ 548.641812] env[61986]: [ 548.642807] env[61986]: ERROR nova.db.main.api [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.642807] env[61986]: result = function(*args, **kwargs) [ 548.642807] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.642807] env[61986]: return func(*args, **kwargs) [ 548.642807] env[61986]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 548.642807] env[61986]: result = fn(*args, **kwargs) [ 548.642807] env[61986]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 548.642807] env[61986]: return f(*args, **kwargs) [ 548.642807] env[61986]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 548.642807] env[61986]: return db.service_get_minimum_version(context, binaries) [ 548.642807] env[61986]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 548.642807] env[61986]: _check_db_access() [ 548.642807] env[61986]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 548.642807] env[61986]: stacktrace = ''.join(traceback.format_stack()) [ 548.642807] env[61986]: [ 548.643222] env[61986]: WARNING nova.objects.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 548.643347] env[61986]: WARNING nova.objects.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Failed to get minimum service version for cell f9811cd0-3c0d-4d0c-bd6c-7affc09d0836 [ 548.643795] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Acquiring lock "singleton_lock" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.643971] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Acquired lock "singleton_lock" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.644234] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Releasing lock "singleton_lock" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.644558] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Full set of CONF: {{(pid=61986) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 548.644699] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ******************************************************************************** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 548.644824] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Configuration options gathered from: {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 548.644957] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 548.645156] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 548.645282] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ================================================================================ {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 548.645490] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] allow_resize_to_same_host = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.645657] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] arq_binding_timeout = 300 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.645861] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] backdoor_port = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.646013] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] backdoor_socket = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.646187] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] block_device_allocate_retries = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.646345] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] block_device_allocate_retries_interval = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.646512] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cert = self.pem {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.646680] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.646925] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute_monitors = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.647113] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] config_dir = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.647284] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] config_drive_format = iso9660 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.647413] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.647571] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] config_source = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.647733] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] console_host = devstack {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.647892] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] control_exchange = nova {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.648061] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cpu_allocation_ratio = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.648246] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] daemon = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.648413] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] debug = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.648567] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] default_access_ip_network_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.648729] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] default_availability_zone = nova {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.648879] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] default_ephemeral_format = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.649047] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] default_green_pool_size = 1000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.649281] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.649440] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] default_schedule_zone = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.649594] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] disk_allocation_ratio = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.649750] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] enable_new_services = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.649925] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] enabled_apis = ['osapi_compute'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.650097] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] enabled_ssl_apis = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.650257] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] flat_injected = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.650412] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] force_config_drive = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.650564] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] force_raw_images = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.650726] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] graceful_shutdown_timeout = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.650881] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] heal_instance_info_cache_interval = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.651106] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] host = cpu-1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.651277] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.651436] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.651591] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.651799] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.651962] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_build_timeout = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.652176] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_delete_interval = 300 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.652387] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_format = [instance: %(uuid)s] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.652600] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_name_template = instance-%08x {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.652721] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_usage_audit = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.652891] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_usage_audit_period = month {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.653066] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.653233] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.653398] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] internal_service_availability_zone = internal {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.653554] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] key = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.653710] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] live_migration_retry_count = 30 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.653874] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_color = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654046] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_config_append = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654215] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654373] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_dir = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654528] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654653] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_options = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654809] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_rotate_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.654970] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_rotate_interval_type = days {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.655149] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] log_rotation_type = none {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.655278] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.655400] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.655564] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.655724] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.655897] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.656076] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] long_rpc_timeout = 1800 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.656236] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] max_concurrent_builds = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.656390] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] max_concurrent_live_migrations = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.656542] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] max_concurrent_snapshots = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.656692] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] max_local_block_devices = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.656858] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] max_logfile_count = 30 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657009] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] max_logfile_size_mb = 200 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657171] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] maximum_instance_delete_attempts = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657336] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metadata_listen = 0.0.0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657503] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metadata_listen_port = 8775 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657667] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metadata_workers = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657828] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] migrate_max_retries = -1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.657994] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] mkisofs_cmd = genisoimage {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.658217] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.658349] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] my_ip = 10.180.1.21 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.658509] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] network_allocate_retries = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.658685] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.658852] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659020] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] osapi_compute_listen_port = 8774 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659193] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] osapi_compute_unique_server_name_scope = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659355] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] osapi_compute_workers = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659513] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] password_length = 12 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659668] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] periodic_enable = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659823] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] periodic_fuzzy_delay = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.659986] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] pointer_model = usbtablet {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.660165] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] preallocate_images = none {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.660322] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] publish_errors = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.660448] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] pybasedir = /opt/stack/nova {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.660602] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ram_allocation_ratio = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.660759] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] rate_limit_burst = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.660922] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] rate_limit_except_level = CRITICAL {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.661093] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] rate_limit_interval = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.661252] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reboot_timeout = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.661407] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reclaim_instance_interval = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.661558] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] record = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.661720] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reimage_timeout_per_gb = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.661881] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] report_interval = 120 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662050] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] rescue_timeout = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662208] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reserved_host_cpus = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662363] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reserved_host_disk_mb = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662516] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reserved_host_memory_mb = 512 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662713] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] reserved_huge_pages = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662824] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] resize_confirm_window = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.662980] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] resize_fs_using_block_device = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.663147] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] resume_guests_state_on_host_boot = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.663310] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.663468] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] rpc_response_timeout = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.663623] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] run_external_periodic_tasks = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.663806] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] running_deleted_instance_action = reap {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.663982] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.664154] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] running_deleted_instance_timeout = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.664309] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler_instance_sync_interval = 120 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.664472] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_down_time = 720 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.664635] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] servicegroup_driver = db {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.664785] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] shell_completion = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.664945] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] shelved_offload_time = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.665110] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] shelved_poll_interval = 3600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.665275] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] shutdown_timeout = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.665433] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] source_is_ipv6 = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.665587] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ssl_only = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.665858] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.666048] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] sync_power_state_interval = 600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.666215] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] sync_power_state_pool_size = 1000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.666380] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] syslog_log_facility = LOG_USER {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.666533] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] tempdir = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.666690] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] timeout_nbd = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.666853] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] transport_url = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.667062] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] update_resources_interval = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.667177] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_cow_images = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.667369] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_eventlog = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.667553] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_journal = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.667717] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_json = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.667873] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_rootwrap_daemon = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.668039] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_stderr = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.668201] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] use_syslog = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.668352] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vcpu_pin_set = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.668514] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plugging_is_fatal = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.668690] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plugging_timeout = 300 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.668890] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] virt_mkfs = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.669070] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] volume_usage_poll_interval = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.669232] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] watch_log_file = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.669397] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] web = /usr/share/spice-html5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 548.669586] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_concurrency.disable_process_locking = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.669871] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.670092] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.670352] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.670548] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.670724] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.670891] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.671090] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.auth_strategy = keystone {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.671260] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.compute_link_prefix = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.671434] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.671607] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.dhcp_domain = novalocal {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.671776] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.enable_instance_password = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.671942] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.glance_link_prefix = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.672161] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.672358] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.672527] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.instance_list_per_project_cells = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.672692] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.list_records_by_skipping_down_cells = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.672856] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.local_metadata_per_cell = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.673036] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.max_limit = 1000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.673210] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.metadata_cache_expiration = 15 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.673384] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.neutron_default_tenant_id = default {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.673638] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.response_validation = warn {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.673857] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.use_neutron_default_nets = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.674058] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.674231] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.674404] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.674578] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.674749] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_dynamic_targets = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.674913] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_jsonfile_path = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.675155] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.675364] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.backend = dogpile.cache.memcached {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.675536] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.backend_argument = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.675709] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.config_prefix = cache.oslo {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.675892] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.dead_timeout = 60.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.676071] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.debug_cache_backend = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.676235] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.enable_retry_client = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.676404] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.enable_socket_keepalive = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.676613] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.enabled = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.676788] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.enforce_fips_mode = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.676956] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.expiration_time = 600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.677159] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.hashclient_retry_attempts = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.677302] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.677465] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_dead_retry = 300 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.677626] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_password = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.677796] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.677967] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.678147] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_pool_maxsize = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.678312] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.678473] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_sasl_enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.678654] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.678825] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.678986] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.memcache_username = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.679209] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.proxies = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.679390] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_db = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.679591] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_password = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.679765] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.679948] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.680136] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_server = localhost:6379 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.680309] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_socket_timeout = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.680491] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.redis_username = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.680677] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.retry_attempts = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.680849] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.retry_delay = 0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.681028] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.socket_keepalive_count = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.681198] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.socket_keepalive_idle = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.681361] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.socket_keepalive_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.681523] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.tls_allowed_ciphers = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.681681] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.tls_cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.681860] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.tls_certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.682078] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.tls_enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.682255] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cache.tls_keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.682429] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.682605] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.auth_type = password {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.682770] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.682950] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.683128] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.683334] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.683522] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.cross_az_attach = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.683691] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.debug = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.683855] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.endpoint_template = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.684034] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.http_retries = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.684206] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.684366] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.684539] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.os_region_name = RegionOne {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.684706] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.684913] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cinder.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.685116] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.685282] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.cpu_dedicated_set = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.685443] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.cpu_shared_set = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.685611] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.image_type_exclude_list = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.685796] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.685958] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.686141] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.686342] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.686522] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.686689] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.resource_provider_association_refresh = 300 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.686864] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.687033] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.shutdown_retry_interval = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.687261] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.687400] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] conductor.workers = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.687598] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] console.allowed_origins = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.687785] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] console.ssl_ciphers = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.687962] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] console.ssl_minimum_version = default {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.688149] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] consoleauth.enforce_session_timeout = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.688317] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] consoleauth.token_ttl = 600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.688491] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.688649] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.688812] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.688990] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.689201] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.689369] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.689534] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.689693] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.689852] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.690018] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.690180] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.region_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.690342] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.690570] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.690759] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.service_type = accelerator {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.690929] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.691101] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.691262] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.691419] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.691599] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.691760] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] cyborg.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.692013] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.backend = sqlalchemy {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.692217] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.connection = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.692389] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.connection_debug = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.692560] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.connection_parameters = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.692723] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.connection_recycle_time = 3600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.692886] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.connection_trace = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.693076] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.db_inc_retry_interval = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.693265] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.db_max_retries = 20 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.693453] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.db_max_retry_interval = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.693620] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.db_retry_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.693786] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.max_overflow = 50 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.693950] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.max_pool_size = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.694148] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.max_retries = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.694294] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.694452] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.mysql_wsrep_sync_wait = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.694646] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.pool_timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.694859] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.retry_interval = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.695040] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.slave_connection = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.695205] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.sqlite_synchronous = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.695367] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] database.use_db_reconnect = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.695547] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.backend = sqlalchemy {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.695718] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.connection = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.695911] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.connection_debug = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.696114] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.connection_parameters = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.696286] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.connection_recycle_time = 3600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.696451] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.connection_trace = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.696615] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.db_inc_retry_interval = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.696778] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.db_max_retries = 20 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.696939] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.db_max_retry_interval = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.697115] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.db_retry_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.697361] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.max_overflow = 50 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.697500] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.max_pool_size = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.697663] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.max_retries = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.697834] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.697995] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.698170] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.pool_timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.698334] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.retry_interval = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.698493] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.slave_connection = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.698677] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] api_database.sqlite_synchronous = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.698876] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] devices.enabled_mdev_types = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.699069] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.699249] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.699414] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ephemeral_storage_encryption.enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.699579] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.699750] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.api_servers = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.699914] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.700094] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.700308] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.700489] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.700649] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.700817] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.debug = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.700982] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.default_trusted_certificate_ids = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.701161] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.enable_certificate_validation = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.701323] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.enable_rbd_download = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.701479] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.701682] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.701870] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.702046] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.702208] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.702373] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.num_retries = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.702542] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.rbd_ceph_conf = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.702705] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.rbd_connect_timeout = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.702872] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.rbd_pool = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.703055] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.rbd_user = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.703285] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.region_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.703466] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.703628] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.703832] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.service_type = image {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.704017] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.704181] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.704343] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.704503] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.704734] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.704912] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.verify_glance_signatures = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.705090] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] glance.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.705265] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] guestfs.debug = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.705433] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] mks.enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.705829] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.706090] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] image_cache.manager_interval = 2400 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.706286] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] image_cache.precache_concurrency = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.706463] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] image_cache.remove_unused_base_images = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.706633] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.706804] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.706983] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] image_cache.subdirectory_name = _base {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.707178] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.api_max_retries = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.707342] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.api_retry_interval = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.707547] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.707770] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.auth_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.707948] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.708122] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.708292] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.708457] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.conductor_group = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.708615] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.708773] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.708933] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.709132] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.709323] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.709486] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.709641] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.709807] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.peer_list = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.709969] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.region_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.710138] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.710303] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.serial_console_state_timeout = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.710465] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.710680] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.service_type = baremetal {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.710849] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.shard = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.711024] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.711188] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.711345] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.711501] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.711684] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.711849] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ironic.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.712082] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.712269] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] key_manager.fixed_key = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.712454] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.712616] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.barbican_api_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.712774] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.barbican_endpoint = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.712947] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.barbican_endpoint_type = public {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.713117] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.barbican_region_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.713315] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.713485] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.713665] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.713848] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.714015] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.714182] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.number_of_retries = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.714346] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.retry_delay = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.714506] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.send_service_user_token = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.714671] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.714915] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.715118] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.verify_ssl = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.715286] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican.verify_ssl_path = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.715455] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.715620] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.auth_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.715819] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.715996] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.716176] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.716378] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.716552] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.716720] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.716882] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] barbican_service_user.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.717064] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.approle_role_id = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.717231] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.approle_secret_id = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.717403] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.kv_mountpoint = secret {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.717570] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.kv_path = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.717749] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.kv_version = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.717933] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.namespace = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.718108] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.root_token_id = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.718270] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.ssl_ca_crt_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.718439] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.timeout = 60.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.718602] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.use_ssl = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.718770] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.718968] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.719210] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.auth_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.719384] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.719546] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.719710] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.719869] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.720041] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.720206] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.720373] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.720577] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.720747] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.720910] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.721081] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.region_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.721247] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.721405] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.721580] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.service_type = identity {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.721743] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.721935] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.722115] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.722277] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.722458] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.722619] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] keystone.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.722823] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.connection_uri = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.722987] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_mode = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.723179] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.723433] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_models = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.723567] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_power_governor_high = performance {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.723763] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.723936] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_power_management = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.724127] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.724295] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.device_detach_attempts = 8 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.724456] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.device_detach_timeout = 20 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.724663] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.disk_cachemodes = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.724844] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.disk_prefix = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.725019] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.enabled_perf_events = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.725186] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.file_backed_memory = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.725351] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.gid_maps = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.725506] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.hw_disk_discard = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.725661] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.hw_machine_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.725857] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_rbd_ceph_conf = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.726072] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.726303] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.726496] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_rbd_glance_store_name = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.726675] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_rbd_pool = rbd {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.726852] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_type = default {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.727015] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.images_volume_group = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.727185] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.inject_key = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.727345] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.inject_partition = -2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.727505] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.inject_password = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.727701] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.iscsi_iface = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.727905] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.iser_use_multipath = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.728090] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.728259] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.728421] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_downtime = 500 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.728579] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.728738] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.728900] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_inbound_addr = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.729114] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.729291] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.729451] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_scheme = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.729621] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_timeout_action = abort {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.729789] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_tunnelled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.729953] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_uri = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.730130] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.live_migration_with_native_tls = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.730292] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.max_queues = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.730461] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.730698] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.730860] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.nfs_mount_options = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.732140] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.732342] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.732517] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.732685] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.732853] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.733070] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.num_pcie_ports = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.733263] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.733435] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.pmem_namespaces = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.733603] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.quobyte_client_cfg = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.733920] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.734112] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.734290] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.734482] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.734651] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rbd_secret_uuid = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.734811] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rbd_user = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.734977] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.735167] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.735330] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rescue_image_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.735489] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rescue_kernel_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.735666] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rescue_ramdisk_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.735890] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.736072] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.rx_queue_size = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.736248] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.smbfs_mount_options = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.736525] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.736698] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.snapshot_compression = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.736865] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.snapshot_image_format = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.737137] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.737326] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.sparse_logical_volumes = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.737494] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.swtpm_enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.737666] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.swtpm_group = tss {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.737836] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.swtpm_user = tss {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.738022] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.sysinfo_serial = unique {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.738189] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.tb_cache_size = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.738349] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.tx_queue_size = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.738537] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.uid_maps = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.738712] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.use_virtio_for_bridges = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.738886] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.virt_type = kvm {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.739072] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.volume_clear = zero {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.739241] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.volume_clear_size = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.739408] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.volume_use_multipath = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.739568] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_cache_path = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.739738] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.739996] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.740199] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.740375] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.740654] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.740834] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.vzstorage_mount_user = stack {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.741014] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.741212] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.741408] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.auth_type = password {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.741574] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.741733] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.741897] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.742067] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.742226] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.742394] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.default_floating_pool = public {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.742556] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.742743] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.extension_sync_interval = 600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.742909] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.http_retries = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.743082] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.743245] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.743401] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.743567] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.743762] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.743987] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.ovs_bridge = br-int {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.744193] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.physnets = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.744367] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.region_name = RegionOne {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.744526] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.744693] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.service_metadata_proxy = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.744850] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.745028] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.service_type = network {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.745194] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.745380] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.745550] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.745710] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.745914] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.746095] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] neutron.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.746272] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] notifications.bdms_in_notifications = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.746449] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] notifications.default_level = INFO {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.746626] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] notifications.notification_format = unversioned {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.746871] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] notifications.notify_on_state_change = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.747106] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.747298] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] pci.alias = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.747474] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] pci.device_spec = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.747638] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] pci.report_in_placement = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.747813] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.747982] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.auth_type = password {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.748208] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.748389] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.748551] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.748714] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.748874] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.749043] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.749205] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.default_domain_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.749362] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.default_domain_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.749548] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.domain_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.749718] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.domain_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.749880] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.750057] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.750219] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.750376] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.750531] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.750698] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.password = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.750861] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.project_domain_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.751089] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.project_domain_name = Default {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.751275] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.project_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.751453] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.project_name = service {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.751624] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.region_name = RegionOne {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.751789] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.751985] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.752179] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.service_type = placement {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.752368] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.752563] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.752730] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.752891] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.system_scope = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.753062] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.753224] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.trust_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.753381] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.user_domain_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.753551] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.user_domain_name = Default {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.753716] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.user_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.753958] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.username = nova {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.754172] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.754340] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] placement.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.754521] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.cores = 20 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.754693] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.count_usage_from_placement = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.754861] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.755043] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.injected_file_content_bytes = 10240 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.755239] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.injected_file_path_length = 255 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.755431] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.injected_files = 5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.755601] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.instances = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.755793] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.key_pairs = 100 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.755961] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.metadata_items = 128 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.756143] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.ram = 51200 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.756308] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.recheck_quota = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.756471] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.server_group_members = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.756677] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] quota.server_groups = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.756880] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.757056] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.757223] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.image_metadata_prefilter = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.757385] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.757546] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.max_attempts = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.757705] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.max_placement_results = 1000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.757869] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.758108] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.758317] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.758505] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] scheduler.workers = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.758692] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.758868] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.759061] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.759238] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.759407] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.759599] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.759790] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.759995] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.760176] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.host_subset_size = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.760344] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.760505] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.760667] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.760833] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.isolated_hosts = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.761027] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.isolated_images = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.761272] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.761453] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.761619] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.761781] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.pci_in_placement = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.761945] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.762121] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.762288] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.762458] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.762661] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.762841] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.763013] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.track_instance_changes = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.763201] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.763369] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metrics.required = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.763532] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metrics.weight_multiplier = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.763717] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.763932] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] metrics.weight_setting = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.764283] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.764473] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] serial_console.enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.764657] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] serial_console.port_range = 10000:20000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.764829] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.765005] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.765182] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] serial_console.serialproxy_port = 6083 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.765360] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.765538] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.auth_type = password {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.765699] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.765887] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.766067] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.766235] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.766394] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.766564] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.send_service_user_token = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.766775] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.766948] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] service_user.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.767135] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.agent_enabled = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.767302] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.767636] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.767832] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.768013] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.html5proxy_port = 6082 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.768235] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.image_compression = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.768404] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.jpeg_compression = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.768565] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.playback_compression = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.768735] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.server_listen = 127.0.0.1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.768903] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.769077] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.streaming_mode = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.769239] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] spice.zlib_compression = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.769417] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] upgrade_levels.baseapi = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.769633] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] upgrade_levels.compute = auto {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.769809] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] upgrade_levels.conductor = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.769969] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] upgrade_levels.scheduler = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.770152] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.770315] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.770473] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.770629] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.770795] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.771008] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.771182] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.771346] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.771503] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vendordata_dynamic_auth.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.771676] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.api_retry_count = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.771837] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.ca_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.772023] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.772193] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.cluster_name = testcl1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.772438] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.connection_pool_size = 10 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.772621] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.console_delay_seconds = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.772795] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.datastore_regex = ^datastore.* {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.773009] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.773196] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.host_password = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.773366] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.host_port = 443 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.773536] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.host_username = administrator@vsphere.local {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.773731] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.insecure = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.773941] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.integration_bridge = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.774132] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.maximum_objects = 100 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.774295] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.pbm_default_policy = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.774456] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.pbm_enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.774613] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.pbm_wsdl_location = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.774785] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.774947] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.serial_port_proxy_uri = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.775124] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.serial_port_service_uri = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.775331] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.task_poll_interval = 0.5 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.775516] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.use_linked_clone = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.775694] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.vnc_keymap = en-us {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.775910] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.vnc_port = 5900 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.776090] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vmware.vnc_port_total = 10000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.776285] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.auth_schemes = ['none'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.776462] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.776785] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.776972] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.777157] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.novncproxy_port = 6080 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.777333] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.server_listen = 127.0.0.1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.777505] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.777664] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.vencrypt_ca_certs = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.777915] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.vencrypt_client_cert = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.778139] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vnc.vencrypt_client_key = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.778328] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.778495] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.disable_deep_image_inspection = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.778659] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.778824] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.778987] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.779165] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.disable_rootwrap = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.779343] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.enable_numa_live_migration = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.779520] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.779690] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.779845] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.780011] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.libvirt_disable_apic = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.780178] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.780339] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.780499] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.780673] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.780876] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.781057] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.781224] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.781385] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.781546] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.781710] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.781899] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.782091] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.client_socket_timeout = 900 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.782295] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.default_pool_size = 1000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.782466] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.keep_alive = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.782633] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.max_header_line = 16384 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.782795] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.782957] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.ssl_ca_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.783129] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.ssl_cert_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.783289] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.ssl_key_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.783452] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.tcp_keepidle = 600 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.783679] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.783867] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] zvm.ca_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.784045] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] zvm.cloud_connector_url = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.784354] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.784529] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] zvm.reachable_timeout = 300 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.784711] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.enforce_new_defaults = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.785156] env[61986]: WARNING oslo_config.cfg [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 548.785360] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.enforce_scope = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.785542] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.policy_default_rule = default {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.785727] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.785928] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.policy_file = policy.yaml {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.786122] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.786287] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.786456] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.786647] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.786816] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.787010] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.787206] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.787384] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.connection_string = messaging:// {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.787552] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.enabled = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.787721] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.es_doc_type = notification {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.787890] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.es_scroll_size = 10000 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.788118] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.es_scroll_time = 2m {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.788343] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.filter_error_trace = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.788480] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.hmac_keys = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.788649] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.sentinel_service_name = mymaster {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.788819] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.socket_timeout = 0.1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.788985] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.trace_requests = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.789162] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler.trace_sqlalchemy = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.789344] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler_jaeger.process_tags = {} {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.789551] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler_jaeger.service_name_prefix = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.789726] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] profiler_otlp.service_name_prefix = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.789892] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] remote_debug.host = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.790066] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] remote_debug.port = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.790252] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.790415] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.790580] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.790744] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.790979] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.791189] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.791359] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.791521] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.791685] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.791861] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.792021] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.792200] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.792385] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.792585] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.792764] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.792935] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.793111] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.793289] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.793451] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.793614] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.793793] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.793995] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.794182] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.794352] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.794515] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.794678] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.794841] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.795010] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.795184] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.795361] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.ssl = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.795538] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.795708] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.795875] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.796055] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.796231] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.796397] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.796588] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.796800] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_notifications.retry = -1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.796997] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.797190] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.797366] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.auth_section = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.797529] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.auth_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.797686] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.cafile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.797843] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.certfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.798022] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.collect_timing = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.798218] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.connect_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.798389] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.connect_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.798547] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.endpoint_id = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.798704] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.endpoint_override = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.798865] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.insecure = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.799037] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.keyfile = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.799200] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.max_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.799355] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.min_version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.799519] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.region_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.799702] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.retriable_status_codes = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.799865] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.service_name = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.800033] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.service_type = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.800202] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.split_loggers = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.800361] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.status_code_retries = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.800519] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.status_code_retry_delay = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.800677] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.timeout = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.800835] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.valid_interfaces = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.801045] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_limit.version = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.801236] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_reports.file_event_handler = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.801406] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.801566] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] oslo_reports.log_dir = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.801740] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.801905] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.802077] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.802255] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.802453] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.802621] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.802797] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.802960] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_ovs_privileged.group = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.803133] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.803302] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.803467] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.803628] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] vif_plug_ovs_privileged.user = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.803825] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.804029] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.804212] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.804411] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.804556] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.804722] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.804890] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.805069] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.805303] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.805494] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.isolate_vif = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.805668] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.805851] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.806047] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.806227] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.806393] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_vif_ovs.per_port_bridge = False {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.806560] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_brick.lock_path = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.806749] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.806943] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.807135] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] privsep_osbrick.capabilities = [21] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.807299] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] privsep_osbrick.group = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.807459] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] privsep_osbrick.helper_command = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.807624] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.807788] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.807948] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] privsep_osbrick.user = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.808135] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.808305] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] nova_sys_admin.group = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.808464] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] nova_sys_admin.helper_command = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.808630] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.808791] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.808948] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] nova_sys_admin.user = None {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 548.809091] env[61986]: DEBUG oslo_service.service [None req-f2f4020b-37b7-4222-b3e9-b89cbd0fcdd4 None None] ******************************************************************************** {{(pid=61986) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 548.809596] env[61986]: INFO nova.service [-] Starting compute node (version 0.1.0) [ 549.313460] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Getting list of instances from cluster (obj){ [ 549.313460] env[61986]: value = "domain-c8" [ 549.313460] env[61986]: _type = "ClusterComputeResource" [ 549.313460] env[61986]: } {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 549.313831] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d864a98-a579-4db3-b95a-00bf3b4b7ef1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.323242] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Got total of 0 instances {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 549.323854] env[61986]: WARNING nova.virt.vmwareapi.driver [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 549.324500] env[61986]: INFO nova.virt.node [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Generated node identity 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d [ 549.324563] env[61986]: INFO nova.virt.node [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Wrote node identity 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d to /opt/stack/data/n-cpu-1/compute_id [ 549.827767] env[61986]: WARNING nova.compute.manager [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Compute nodes ['2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 550.833227] env[61986]: INFO nova.compute.manager [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 551.838736] env[61986]: WARNING nova.compute.manager [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 551.839143] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.839226] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.839312] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.839465] env[61986]: DEBUG nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 551.840406] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373893c6-fb07-4831-b61f-b05d45a89309 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.848642] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f1a5a5-7768-442b-a64f-0416aead5f46 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.861957] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ba5488-2584-4df5-8647-707f246e98fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.868229] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5a1e1d-e224-4ecd-9642-836088a24da0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.897251] env[61986]: DEBUG nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181536MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 551.897446] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.897572] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.400509] env[61986]: WARNING nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] No compute node record for cpu-1:2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d could not be found. [ 552.904734] env[61986]: INFO nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d [ 554.416158] env[61986]: DEBUG nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 554.416158] env[61986]: DEBUG nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 554.580362] env[61986]: INFO nova.scheduler.client.report [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] [req-48d79ca8-8585-46e8-a329-800ec33869b4] Created resource provider record via placement API for resource provider with UUID 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 554.595845] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620c7d56-3cdc-4004-9c95-f3d5ad172053 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.604076] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4e8e50-768f-4ccd-89bd-2bca651e7305 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.634776] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80d40cd-ef22-4d7f-aba9-acd803bc4b49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.642216] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6114ee4f-6a2b-4ac7-bb9a-af456c516406 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.656639] env[61986]: DEBUG nova.compute.provider_tree [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 555.192987] env[61986]: DEBUG nova.scheduler.client.report [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 555.192987] env[61986]: DEBUG nova.compute.provider_tree [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 0 to 1 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 555.192987] env[61986]: DEBUG nova.compute.provider_tree [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 555.247026] env[61986]: DEBUG nova.compute.provider_tree [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 1 to 2 during operation: update_traits {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 555.752102] env[61986]: DEBUG nova.compute.resource_tracker [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 555.752102] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.852s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.752102] env[61986]: DEBUG nova.service [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Creating RPC server for service compute {{(pid=61986) start /opt/stack/nova/nova/service.py:186}} [ 555.765656] env[61986]: DEBUG nova.service [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] Join ServiceGroup membership for this service compute {{(pid=61986) start /opt/stack/nova/nova/service.py:203}} [ 555.766298] env[61986]: DEBUG nova.servicegroup.drivers.db [None req-df16a314-d72a-4714-9974-0a3d7af1ccca None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61986) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 592.682418] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquiring lock "69e3c117-ee31-4d80-812b-da3c12fe19b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.682418] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Lock "69e3c117-ee31-4d80-812b-da3c12fe19b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.190081] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.605587] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "f7f3a069-04da-4090-bbb1-1406799b5fdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.605828] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "f7f3a069-04da-4090-bbb1-1406799b5fdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.738719] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.738969] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.744163] env[61986]: INFO nova.compute.claims [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.109318] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 594.422377] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "6b586395-eb84-44f6-84c8-4a3ef3d64254" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.422664] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "6b586395-eb84-44f6-84c8-4a3ef3d64254" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.640102] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.840456] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ce2b2f-abf9-43eb-8bcf-7d3d4e401c0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.848560] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6a4801-2285-4417-9dff-e18c7ac20dcc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.882764] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee361594-934d-4e60-b82e-d48e93868ad7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.891812] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c1e825-ce7d-4e68-9553-0e70678280be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.906866] env[61986]: DEBUG nova.compute.provider_tree [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.935209] env[61986]: DEBUG nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.050746] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.051928] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.414012] env[61986]: DEBUG nova.scheduler.client.report [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.463853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.554691] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.910360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquiring lock "bb58f75b-97d0-4615-a014-3de3bf7dee2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.911638] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Lock "bb58f75b-97d0-4615-a014-3de3bf7dee2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.924885] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.185s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.924885] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.930745] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.291s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.932247] env[61986]: INFO nova.compute.claims [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.085310] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.418290] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.434031] env[61986]: DEBUG nova.compute.utils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.435864] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.435864] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.768675] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_power_states {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.817025] env[61986]: DEBUG nova.policy [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '068bc72c6fb64bf28683069c393505e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d449e6314c24967b8bc6d6636d38a3f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 596.943493] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.949897] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.047277] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a69f84-571e-4281-91d0-bd07e18892ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.058104] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1ef422-8e3b-4c11-8042-0fe2995c0f4b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.091237] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4e23f5-4582-4f0b-9781-194687b1f343 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.101337] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb485dd4-aa22-4829-99c0-56415a9aec58 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.115914] env[61986]: DEBUG nova.compute.provider_tree [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.274423] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Getting list of instances from cluster (obj){ [ 597.274423] env[61986]: value = "domain-c8" [ 597.274423] env[61986]: _type = "ClusterComputeResource" [ 597.274423] env[61986]: } {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 597.275498] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e4ecc0-5b84-4c48-a201-139af0fded24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.286067] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Got total of 0 instances {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 597.286309] env[61986]: WARNING nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] While synchronizing instance power states, found 2 instances in the database and 0 instances on the hypervisor. [ 597.286388] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid 69e3c117-ee31-4d80-812b-da3c12fe19b9 {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 597.286587] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid f7f3a069-04da-4090-bbb1-1406799b5fdd {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 597.286960] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "69e3c117-ee31-4d80-812b-da3c12fe19b9" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.287801] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "f7f3a069-04da-4090-bbb1-1406799b5fdd" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.287801] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.287938] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Getting list of instances from cluster (obj){ [ 597.287938] env[61986]: value = "domain-c8" [ 597.287938] env[61986]: _type = "ClusterComputeResource" [ 597.287938] env[61986]: } {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 597.288829] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84cd23d5-9b30-4400-a82e-d8ccb6ada94b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.296826] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Got total of 0 instances {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 597.373684] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Successfully created port: 1c0b8587-acc7-4d94-860c-e33e46a2e1ac {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.622335] env[61986]: DEBUG nova.scheduler.client.report [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.962263] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.993069] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.993359] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.993525] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.993800] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.993999] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.994216] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.994444] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.994595] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.994975] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.995158] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.995326] env[61986]: DEBUG nova.virt.hardware [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.996310] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfdf357-fd4d-49ea-9c5a-c9bdb36a0ef5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.007066] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb05d8e-42c6-4053-a9b3-e4137d8582b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.024232] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c995f8-77b7-4988-8968-c59346d966a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.129761] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.130651] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.134515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.671s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.136452] env[61986]: INFO nova.compute.claims [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.636254] env[61986]: DEBUG nova.compute.utils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.638016] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.643700] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 598.759259] env[61986]: DEBUG nova.policy [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48052f2800ad4caea81169b99fa55d1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'def2168f736745bba50861f9800a4c6b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 599.151458] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.266084] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a46c198-2719-49fe-b92a-45daa24481d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.277083] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc863ff3-b399-4c74-b23f-9c71ab57f75f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.313525] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684204a3-62c9-4d0b-9e9e-f956ef2df9c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.321711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bfa115-6896-4ffe-9675-7d292e3932e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.335546] env[61986]: DEBUG nova.compute.provider_tree [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.475888] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Successfully created port: 6395ca6c-109f-419c-93a1-5552d949daa1 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.626183] env[61986]: ERROR nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 599.626183] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.626183] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.626183] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.626183] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.626183] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.626183] env[61986]: ERROR nova.compute.manager raise self.value [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.626183] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.626183] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.626183] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.626598] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.626598] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.626598] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 599.626598] env[61986]: ERROR nova.compute.manager [ 599.626598] env[61986]: Traceback (most recent call last): [ 599.626598] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.626598] env[61986]: listener.cb(fileno) [ 599.626598] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.626598] env[61986]: result = function(*args, **kwargs) [ 599.626598] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.626598] env[61986]: return func(*args, **kwargs) [ 599.626598] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.626598] env[61986]: raise e [ 599.626598] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.626598] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 599.626598] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.626598] env[61986]: created_port_ids = self._update_ports_for_instance( [ 599.626598] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.626598] env[61986]: with excutils.save_and_reraise_exception(): [ 599.626598] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.626598] env[61986]: self.force_reraise() [ 599.626598] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.626598] env[61986]: raise self.value [ 599.626598] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.626598] env[61986]: updated_port = self._update_port( [ 599.626598] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.626598] env[61986]: _ensure_no_port_binding_failure(port) [ 599.626598] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.626598] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.627317] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 599.627317] env[61986]: Removing descriptor: 14 [ 599.628257] env[61986]: ERROR nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Traceback (most recent call last): [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] yield resources [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self.driver.spawn(context, instance, image_meta, [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] vm_ref = self.build_virtual_machine(instance, [ 599.628257] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] for vif in network_info: [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return self._sync_wrapper(fn, *args, **kwargs) [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self.wait() [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self[:] = self._gt.wait() [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return self._exit_event.wait() [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.628669] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] result = hub.switch() [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return self.greenlet.switch() [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] result = function(*args, **kwargs) [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return func(*args, **kwargs) [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] raise e [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] nwinfo = self.network_api.allocate_for_instance( [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] created_port_ids = self._update_ports_for_instance( [ 599.629012] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] with excutils.save_and_reraise_exception(): [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self.force_reraise() [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] raise self.value [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] updated_port = self._update_port( [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] _ensure_no_port_binding_failure(port) [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] raise exception.PortBindingFailed(port_id=port['id']) [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 599.629349] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] [ 599.629695] env[61986]: INFO nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Terminating instance [ 599.631890] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquiring lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.632121] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquired lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.632415] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.839430] env[61986]: DEBUG nova.scheduler.client.report [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.167197] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.208986] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.208986] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.208986] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.209229] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.210246] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.210438] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.210647] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.210831] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.210995] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.211169] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.211332] env[61986]: DEBUG nova.virt.hardware [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.212230] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27f81c7-124f-4a39-98b7-18b2cef66749 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.215955] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.227435] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c87d5c-fa6c-484a-9de5-fff2c1559c35 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.319659] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.349374] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.215s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.349773] env[61986]: DEBUG nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.356342] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.268s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.356342] env[61986]: INFO nova.compute.claims [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.615126] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "03489669-961b-4e5b-87b1-ef05bb59433b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.615512] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "03489669-961b-4e5b-87b1-ef05bb59433b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.823703] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Releasing lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.824166] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 600.824354] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 600.824654] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8a1f5aa-b632-40f1-9ed2-9c0d97d66906 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.837172] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5afda0-020f-46ce-9af4-64270bd7aef2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.858692] env[61986]: DEBUG nova.compute.utils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.863047] env[61986]: DEBUG nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 600.880281] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69e3c117-ee31-4d80-812b-da3c12fe19b9 could not be found. [ 600.880509] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 600.880896] env[61986]: INFO nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Took 0.06 seconds to destroy the instance on the hypervisor. [ 600.881169] env[61986]: DEBUG oslo.service.loopingcall [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.882271] env[61986]: DEBUG nova.compute.manager [-] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.882271] env[61986]: DEBUG nova.network.neutron [-] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.943988] env[61986]: DEBUG nova.network.neutron [-] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.118401] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.129355] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "76b7db85-734f-40fe-8d58-2fba89bf205f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.129510] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "76b7db85-734f-40fe-8d58-2fba89bf205f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.364297] env[61986]: DEBUG nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.396922] env[61986]: DEBUG nova.compute.manager [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Received event network-changed-1c0b8587-acc7-4d94-860c-e33e46a2e1ac {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 601.397265] env[61986]: DEBUG nova.compute.manager [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Refreshing instance network info cache due to event network-changed-1c0b8587-acc7-4d94-860c-e33e46a2e1ac. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 601.397585] env[61986]: DEBUG oslo_concurrency.lockutils [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] Acquiring lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.397869] env[61986]: DEBUG oslo_concurrency.lockutils [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] Acquired lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.398229] env[61986]: DEBUG nova.network.neutron [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Refreshing network info cache for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 601.447694] env[61986]: DEBUG nova.network.neutron [-] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.469427] env[61986]: ERROR nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 601.469427] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.469427] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.469427] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.469427] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.469427] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.469427] env[61986]: ERROR nova.compute.manager raise self.value [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.469427] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.469427] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.469427] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.469917] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.469917] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.469917] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 601.469917] env[61986]: ERROR nova.compute.manager [ 601.469917] env[61986]: Traceback (most recent call last): [ 601.469917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.469917] env[61986]: listener.cb(fileno) [ 601.469917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.469917] env[61986]: result = function(*args, **kwargs) [ 601.469917] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.469917] env[61986]: return func(*args, **kwargs) [ 601.469917] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.469917] env[61986]: raise e [ 601.469917] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.469917] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 601.469917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.469917] env[61986]: created_port_ids = self._update_ports_for_instance( [ 601.469917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.469917] env[61986]: with excutils.save_and_reraise_exception(): [ 601.469917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.469917] env[61986]: self.force_reraise() [ 601.469917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.469917] env[61986]: raise self.value [ 601.469917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.469917] env[61986]: updated_port = self._update_port( [ 601.469917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.469917] env[61986]: _ensure_no_port_binding_failure(port) [ 601.469917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.469917] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.470668] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 601.470668] env[61986]: Removing descriptor: 15 [ 601.470668] env[61986]: ERROR nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Traceback (most recent call last): [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] yield resources [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self.driver.spawn(context, instance, image_meta, [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.470668] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] vm_ref = self.build_virtual_machine(instance, [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] for vif in network_info: [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return self._sync_wrapper(fn, *args, **kwargs) [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self.wait() [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self[:] = self._gt.wait() [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return self._exit_event.wait() [ 601.470952] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] result = hub.switch() [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return self.greenlet.switch() [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] result = function(*args, **kwargs) [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return func(*args, **kwargs) [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] raise e [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] nwinfo = self.network_api.allocate_for_instance( [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.471267] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] created_port_ids = self._update_ports_for_instance( [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] with excutils.save_and_reraise_exception(): [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self.force_reraise() [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] raise self.value [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] updated_port = self._update_port( [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] _ensure_no_port_binding_failure(port) [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.471592] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] raise exception.PortBindingFailed(port_id=port['id']) [ 601.471872] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 601.471872] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] [ 601.471872] env[61986]: INFO nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Terminating instance [ 601.478895] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.479119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquired lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.479238] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.523268] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8635e5-79ee-4138-8374-5a6b9f7fe487 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.532842] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc952b10-8a43-465a-ade5-8e000deda37f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.570306] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83652221-6ee3-43de-9420-383e0385e36d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.578027] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cffb70-c572-414c-be7b-cd6cf63bc275 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.591411] env[61986]: DEBUG nova.compute.provider_tree [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.635279] env[61986]: DEBUG nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.648097] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.941319] env[61986]: DEBUG nova.network.neutron [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.953515] env[61986]: INFO nova.compute.manager [-] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Took 1.07 seconds to deallocate network for instance. [ 601.956252] env[61986]: DEBUG nova.compute.claims [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 601.957090] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.003998] env[61986]: DEBUG nova.network.neutron [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.005775] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.094853] env[61986]: DEBUG nova.scheduler.client.report [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.132365] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.161285] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.374069] env[61986]: DEBUG nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.403413] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.403652] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.403806] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.403978] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.404138] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.404348] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.404584] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.404642] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.404782] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.405611] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.405853] env[61986]: DEBUG nova.virt.hardware [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.407013] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88d7cda-9858-46d8-be95-410ef1bfe994 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.418030] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c7fb23-9698-4c94-bb3d-da8e0765d4c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.432362] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 602.444753] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 602.445066] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f2d4fd5-f084-4b5a-8b98-55539ec12f68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.457848] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Created folder: OpenStack in parent group-v4. [ 602.457952] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Creating folder: Project (1eb9e509c9544e57a32cbcf8a90720e1). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 602.458188] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f97f1a2a-3650-4cc1-8229-e05ab2d853ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.468906] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Created folder: Project (1eb9e509c9544e57a32cbcf8a90720e1) in parent group-v252271. [ 602.469134] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Creating folder: Instances. Parent ref: group-v252272. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 602.469928] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b734b9b-4601-4a72-aad0-a85f55fb3b71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.480674] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Created folder: Instances in parent group-v252272. [ 602.480911] env[61986]: DEBUG oslo.service.loopingcall [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.481613] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 602.481850] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69810c47-48e2-4981-b2fd-77f33622506d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.500764] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 602.500764] env[61986]: value = "task-1159490" [ 602.500764] env[61986]: _type = "Task" [ 602.500764] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.511403] env[61986]: DEBUG oslo_concurrency.lockutils [req-4f043935-6e50-451e-aa38-b8b1680805a1 req-a5e2b733-9e29-4266-b53b-350a5e7a927e service nova] Releasing lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.511720] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159490, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.599799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.600335] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.606018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.653s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.606018] env[61986]: INFO nova.compute.claims [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.636961] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Releasing lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.638343] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.638343] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 602.638343] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ca62199-9f22-40c9-89d9-c1d1b29de1b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.646746] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa7bc22-ebad-4630-9437-98ac11302d61 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.670951] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f7f3a069-04da-4090-bbb1-1406799b5fdd could not be found. [ 602.671269] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 602.671457] env[61986]: INFO nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 602.671694] env[61986]: DEBUG oslo.service.loopingcall [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.671898] env[61986]: DEBUG nova.compute.manager [-] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.671996] env[61986]: DEBUG nova.network.neutron [-] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.706574] env[61986]: DEBUG nova.network.neutron [-] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.011482] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159490, 'name': CreateVM_Task, 'duration_secs': 0.335099} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.012376] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 603.013773] env[61986]: DEBUG oslo_vmware.service [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444fb871-060c-400d-9728-34a07670cdd0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.020962] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.021202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.021984] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 603.022186] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41688d31-8d0a-491d-b488-db5aebb61f4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.027616] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 603.027616] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e28380-7701-2e78-b379-6c3badf40135" [ 603.027616] env[61986]: _type = "Task" [ 603.027616] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.038523] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e28380-7701-2e78-b379-6c3badf40135, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.110327] env[61986]: DEBUG nova.compute.utils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.117647] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.117854] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 603.211473] env[61986]: DEBUG nova.network.neutron [-] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.257381] env[61986]: DEBUG nova.policy [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6881bdd48c4f0ea6bcb7d2d3746d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e9d0c35a4de4f5f9829a7f3c88fde92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 603.540923] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.545924] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 603.545924] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.545924] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.545924] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 603.545924] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9397b90-78af-498a-a718-efef17fcea9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.562057] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 603.562135] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 603.563386] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a361702-919e-4ab8-8235-a6dd683125b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.571193] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30447da0-904a-4337-bede-71add97ed48c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.577256] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 603.577256] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5215ff74-d62d-b9ea-1042-4b26b4de1b9d" [ 603.577256] env[61986]: _type = "Task" [ 603.577256] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.585724] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5215ff74-d62d-b9ea-1042-4b26b4de1b9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.618533] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.716564] env[61986]: INFO nova.compute.manager [-] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Took 1.04 seconds to deallocate network for instance. [ 603.725029] env[61986]: DEBUG nova.compute.claims [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 603.725029] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.758949] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb2fd4d-ab3b-4c83-b240-6cb31ed6e36a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.769294] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b54f4e-060b-4420-984b-4e8d5e08f541 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.802279] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11467664-ed54-49c8-bdb8-29777e4595ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.813879] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665ca1dc-c3c5-491f-a2c6-34cc6686b4d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.825730] env[61986]: DEBUG nova.compute.provider_tree [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.854637] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Successfully created port: a9a9f75e-f803-4732-b2e3-7aa9e414daf9 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.881673] env[61986]: DEBUG nova.compute.manager [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Received event network-changed-6395ca6c-109f-419c-93a1-5552d949daa1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.881884] env[61986]: DEBUG nova.compute.manager [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Refreshing instance network info cache due to event network-changed-6395ca6c-109f-419c-93a1-5552d949daa1. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 603.882395] env[61986]: DEBUG oslo_concurrency.lockutils [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] Acquiring lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.882530] env[61986]: DEBUG oslo_concurrency.lockutils [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] Acquired lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.882696] env[61986]: DEBUG nova.network.neutron [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Refreshing network info cache for port 6395ca6c-109f-419c-93a1-5552d949daa1 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.090274] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 604.090560] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Creating directory with path [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 604.090802] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cacb4948-1bd8-441a-8963-75623b97ea97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.124888] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Created directory with path [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 604.125111] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Fetch image to [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 604.125278] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Downloading image file data 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk on the data store datastore2 {{(pid=61986) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 604.126197] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3edd9524-5548-4b56-a350-6168154a418b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.145949] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08331fda-5d2e-4d21-9903-eba291dfdc27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.160978] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91073d2-5c15-4701-b63e-f6005b3f0555 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.203089] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a00b381-d812-4a31-8096-e2e28eadb112 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.210740] env[61986]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-13174612-4256-4f3c-b363-37b0b95d7e68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.224427] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.224427] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.225195] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 604.225195] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Rebuilding the list of instances to heal {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 604.299076] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Downloading image file data 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to the data store datastore2 {{(pid=61986) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 604.328482] env[61986]: DEBUG nova.scheduler.client.report [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.385732] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 604.452263] env[61986]: DEBUG nova.network.neutron [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.519149] env[61986]: DEBUG nova.network.neutron [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.646776] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.678171] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.678937] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.678937] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.678937] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.678937] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.679238] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.679238] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.679389] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.679612] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.679694] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.679854] env[61986]: DEBUG nova.virt.hardware [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.680755] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930ec829-788c-4bd3-8b00-4ca6cd5ae1c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.691011] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240b7a2e-a631-4e04-954a-7686961e2ade {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.729786] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 604.729886] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 604.730016] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 604.730091] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 604.730774] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 604.732042] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Didn't find any instances for network info cache update. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 604.736208] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.736208] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.736208] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.736422] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.736779] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.736974] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.736974] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 604.737236] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 604.840272] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.840478] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.844425] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.196s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.846372] env[61986]: INFO nova.compute.claims [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.025324] env[61986]: DEBUG oslo_concurrency.lockutils [req-b102657b-f716-4040-a2fd-9023026c68ce req-66e3f583-5bda-44e3-a7bc-e59212382b1e service nova] Releasing lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.071028] env[61986]: DEBUG nova.compute.manager [req-cda21c14-250a-4540-a26c-45495ed09e9c req-f8c51781-dab4-4d55-a5f7-35e8a579097d service nova] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Received event network-vif-deleted-1c0b8587-acc7-4d94-860c-e33e46a2e1ac {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.171356] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 605.171621] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 605.222347] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Downloaded image file data 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk on the data store datastore2 {{(pid=61986) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 605.224331] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 605.224444] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Copying Virtual Disk [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk to [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 605.224697] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c53fa9e0-03f1-4da5-ae47-9a5f6f98470f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.236755] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 605.236755] env[61986]: value = "task-1159491" [ 605.236755] env[61986]: _type = "Task" [ 605.236755] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.246194] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.247875] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.350347] env[61986]: DEBUG nova.compute.utils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.351845] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.352134] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 605.488261] env[61986]: DEBUG nova.policy [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a0e6962135de435ba0ae2ae1fe10cb97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55d086833bd348db82cae2e6460429b1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 605.696673] env[61986]: ERROR nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 605.696673] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.696673] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.696673] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.696673] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.696673] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.696673] env[61986]: ERROR nova.compute.manager raise self.value [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.696673] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.696673] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.696673] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.697438] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.697438] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.697438] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 605.697438] env[61986]: ERROR nova.compute.manager [ 605.697438] env[61986]: Traceback (most recent call last): [ 605.697438] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.697438] env[61986]: listener.cb(fileno) [ 605.697438] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.697438] env[61986]: result = function(*args, **kwargs) [ 605.697438] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.697438] env[61986]: return func(*args, **kwargs) [ 605.697438] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.697438] env[61986]: raise e [ 605.697438] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.697438] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 605.697438] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.697438] env[61986]: created_port_ids = self._update_ports_for_instance( [ 605.697438] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.697438] env[61986]: with excutils.save_and_reraise_exception(): [ 605.697438] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.697438] env[61986]: self.force_reraise() [ 605.697438] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.697438] env[61986]: raise self.value [ 605.697438] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.697438] env[61986]: updated_port = self._update_port( [ 605.697438] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.697438] env[61986]: _ensure_no_port_binding_failure(port) [ 605.697438] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.697438] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.698276] env[61986]: nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 605.698276] env[61986]: Removing descriptor: 14 [ 605.698276] env[61986]: ERROR nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Traceback (most recent call last): [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] yield resources [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self.driver.spawn(context, instance, image_meta, [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.698276] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] vm_ref = self.build_virtual_machine(instance, [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] for vif in network_info: [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return self._sync_wrapper(fn, *args, **kwargs) [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self.wait() [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self[:] = self._gt.wait() [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return self._exit_event.wait() [ 605.698621] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] result = hub.switch() [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return self.greenlet.switch() [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] result = function(*args, **kwargs) [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return func(*args, **kwargs) [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] raise e [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] nwinfo = self.network_api.allocate_for_instance( [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.699040] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] created_port_ids = self._update_ports_for_instance( [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] with excutils.save_and_reraise_exception(): [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self.force_reraise() [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] raise self.value [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] updated_port = self._update_port( [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] _ensure_no_port_binding_failure(port) [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.699423] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] raise exception.PortBindingFailed(port_id=port['id']) [ 605.699776] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 605.699776] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] [ 605.699776] env[61986]: INFO nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Terminating instance [ 605.702776] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.702776] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.702776] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.750743] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159491, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.858150] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.002658] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf5a831-ee4b-4478-abbe-a6e37ce22ad5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.010798] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec10600-d974-407a-bb41-f02776a76799 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.050255] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4fce6a-e42d-4619-828a-3c733235d2be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.058480] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18876f3-df38-4385-9d06-622aac36609c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.072437] env[61986]: DEBUG nova.compute.provider_tree [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 606.081652] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Successfully created port: 5b6cbc4a-c419-450e-afdf-da4251c8bb11 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.246813] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159491, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66028} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.247065] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Copied Virtual Disk [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk to [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 606.247246] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleting the datastore file [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 606.247487] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef0bfa87-f6c4-4b3b-bc59-1b5dc7ddd940 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.249683] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.257105] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 606.257105] env[61986]: value = "task-1159492" [ 606.257105] env[61986]: _type = "Task" [ 606.257105] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.264381] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.462112] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.601618] env[61986]: ERROR nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [req-f6590769-e475-481c-b34e-31f220aed71b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f6590769-e475-481c-b34e-31f220aed71b"}]} [ 606.626654] env[61986]: DEBUG nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Refreshing inventories for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 606.649613] env[61986]: DEBUG nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updating ProviderTree inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 606.649829] env[61986]: DEBUG nova.compute.provider_tree [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 606.673850] env[61986]: DEBUG nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Refreshing aggregate associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, aggregates: None {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 606.697049] env[61986]: DEBUG nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Refreshing trait associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 606.772119] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027256} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.772818] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 606.773299] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Moving file from [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5. {{(pid=61986) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 606.773960] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-780d6428-65af-46c2-9d4e-c3747003375a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.784046] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 606.784046] env[61986]: value = "task-1159493" [ 606.784046] env[61986]: _type = "Task" [ 606.784046] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.794891] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159493, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.841728] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd79503e-553e-4de5-a9b0-454eaf752b0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.852637] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361f09c6-24c3-42e4-b0e5-dde9607973da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.894507] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.897547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b42cf55-e18f-4c75-bef8-8dcbfa71824d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.906482] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b07e7fa-f43e-4db1-b8dd-0e2e44aacc64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.925596] env[61986]: DEBUG nova.compute.provider_tree [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 606.930694] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.930694] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.930694] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.930946] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.930946] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.930946] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.930946] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.930946] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.931203] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.932335] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.932576] env[61986]: DEBUG nova.virt.hardware [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.933433] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effbc3f2-1567-458c-989c-4325a343db06 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.942719] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd775d0d-aeb7-4890-a152-fa8b9eb08caf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.960447] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "a048d234-fb8d-4904-a016-2d1c6bd6d103" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.960674] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "a048d234-fb8d-4904-a016-2d1c6bd6d103" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.969104] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.969608] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.970668] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 606.970668] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10fc9a72-5e52-4ac6-a869-2f1d64997941 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.980759] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9c60d9-b339-4189-92e6-520ff3e69a07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.003742] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a could not be found. [ 607.003875] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 607.004046] env[61986]: INFO nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 607.004299] env[61986]: DEBUG oslo.service.loopingcall [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.004789] env[61986]: DEBUG nova.compute.manager [-] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.004789] env[61986]: DEBUG nova.network.neutron [-] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 607.040970] env[61986]: DEBUG nova.network.neutron [-] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.297356] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159493, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.027902} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.298622] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] File moved {{(pid=61986) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 607.299024] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Cleaning up location [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 607.299239] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleting the datastore file [datastore2] vmware_temp/78e05b41-8761-4dc0-ac76-77a4a05b6d86 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 607.300420] env[61986]: DEBUG nova.compute.manager [req-9d4a8919-fbbe-4e7d-ad26-756792452c02 req-f5828b8d-af1c-4136-ba3f-d89d6fbaf7fb service nova] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Received event network-vif-deleted-6395ca6c-109f-419c-93a1-5552d949daa1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 607.300797] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b11ce92-a5ae-42b1-8202-02f902aedcbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.309036] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 607.309036] env[61986]: value = "task-1159494" [ 607.309036] env[61986]: _type = "Task" [ 607.309036] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.320068] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159494, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.416251] env[61986]: ERROR nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 607.416251] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.416251] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.416251] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.416251] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.416251] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.416251] env[61986]: ERROR nova.compute.manager raise self.value [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.416251] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.416251] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.416251] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.416710] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.416710] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.416710] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 607.416710] env[61986]: ERROR nova.compute.manager [ 607.416710] env[61986]: Traceback (most recent call last): [ 607.416710] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.416710] env[61986]: listener.cb(fileno) [ 607.416710] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.416710] env[61986]: result = function(*args, **kwargs) [ 607.416710] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.416710] env[61986]: return func(*args, **kwargs) [ 607.416710] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.416710] env[61986]: raise e [ 607.416710] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.416710] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 607.416710] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.416710] env[61986]: created_port_ids = self._update_ports_for_instance( [ 607.416710] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.416710] env[61986]: with excutils.save_and_reraise_exception(): [ 607.416710] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.416710] env[61986]: self.force_reraise() [ 607.416710] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.416710] env[61986]: raise self.value [ 607.416710] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.416710] env[61986]: updated_port = self._update_port( [ 607.416710] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.416710] env[61986]: _ensure_no_port_binding_failure(port) [ 607.416710] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.416710] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.417481] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 607.417481] env[61986]: Removing descriptor: 15 [ 607.417481] env[61986]: ERROR nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Traceback (most recent call last): [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] yield resources [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self.driver.spawn(context, instance, image_meta, [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.417481] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] vm_ref = self.build_virtual_machine(instance, [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] for vif in network_info: [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return self._sync_wrapper(fn, *args, **kwargs) [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self.wait() [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self[:] = self._gt.wait() [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return self._exit_event.wait() [ 607.417852] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] result = hub.switch() [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return self.greenlet.switch() [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] result = function(*args, **kwargs) [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return func(*args, **kwargs) [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] raise e [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] nwinfo = self.network_api.allocate_for_instance( [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.418457] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] created_port_ids = self._update_ports_for_instance( [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] with excutils.save_and_reraise_exception(): [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self.force_reraise() [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] raise self.value [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] updated_port = self._update_port( [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] _ensure_no_port_binding_failure(port) [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.419050] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] raise exception.PortBindingFailed(port_id=port['id']) [ 607.419388] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 607.419388] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] [ 607.419388] env[61986]: INFO nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Terminating instance [ 607.420714] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquiring lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.422031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquired lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.422031] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.463171] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.482343] env[61986]: DEBUG nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 14 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 607.482589] env[61986]: DEBUG nova.compute.provider_tree [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 14 to 15 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 607.482762] env[61986]: DEBUG nova.compute.provider_tree [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 607.544354] env[61986]: DEBUG nova.network.neutron [-] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.819465] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159494, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025792} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.819999] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 607.820658] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac26e9ac-927c-4bc4-ae42-51dc6cdc4ec0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.826139] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 607.826139] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527af99a-2215-6eee-6ad7-954e4357a8bc" [ 607.826139] env[61986]: _type = "Task" [ 607.826139] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.833324] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527af99a-2215-6eee-6ad7-954e4357a8bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.954745] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.987996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.143s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.988561] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.992513] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.992788] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 6.036s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.008188] env[61986]: DEBUG nova.compute.manager [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Received event network-changed-a9a9f75e-f803-4732-b2e3-7aa9e414daf9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.009377] env[61986]: DEBUG nova.compute.manager [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Refreshing instance network info cache due to event network-changed-a9a9f75e-f803-4732-b2e3-7aa9e414daf9. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 608.009377] env[61986]: DEBUG oslo_concurrency.lockutils [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] Acquiring lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.009377] env[61986]: DEBUG oslo_concurrency.lockutils [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] Acquired lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.009377] env[61986]: DEBUG nova.network.neutron [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Refreshing network info cache for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 608.047325] env[61986]: INFO nova.compute.manager [-] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Took 1.04 seconds to deallocate network for instance. [ 608.051364] env[61986]: DEBUG nova.compute.claims [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 608.051364] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.142473] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.338988] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527af99a-2215-6eee-6ad7-954e4357a8bc, 'name': SearchDatastore_Task, 'duration_secs': 0.007881} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.339285] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.339578] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6b586395-eb84-44f6-84c8-4a3ef3d64254/6b586395-eb84-44f6-84c8-4a3ef3d64254.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 608.339833] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f9dc1a7-20a2-4f74-9b43-2803ae1270b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.346843] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 608.346843] env[61986]: value = "task-1159495" [ 608.346843] env[61986]: _type = "Task" [ 608.346843] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.356927] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159495, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.497856] env[61986]: DEBUG nova.compute.utils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.503989] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.504208] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 608.556354] env[61986]: DEBUG nova.network.neutron [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.585540] env[61986]: DEBUG nova.policy [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '415b0e7dfa1541bfba60c687002c440c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5163e73250e14ba89e003d97bdf39ba2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 608.651208] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Releasing lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.651756] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.651976] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 608.658902] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-228cc4ed-9bf5-4301-9ddc-075c3d6fca22 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.660208] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquiring lock "ab392d78-a11d-44d0-81d6-8afcb169e6e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.660208] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Lock "ab392d78-a11d-44d0-81d6-8afcb169e6e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.673623] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fdb1cb-bdd6-41e1-95c4-288bad0c30e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.703506] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bb58f75b-97d0-4615-a014-3de3bf7dee2f could not be found. [ 608.703834] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 608.704067] env[61986]: INFO nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 608.704342] env[61986]: DEBUG oslo.service.loopingcall [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.706233] env[61986]: DEBUG nova.compute.manager [-] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.706233] env[61986]: DEBUG nova.network.neutron [-] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.709045] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b2dc82-e829-49d1-8eb7-3e8238241437 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.718859] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a56209-dcd2-4ff8-b2ea-2d6b150f08da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.759882] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad39e32-a958-4b81-8c25-90b9df9f3c04 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.764200] env[61986]: DEBUG nova.network.neutron [-] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.779185] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2053b254-b025-4777-9b6a-93559f53cd0e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.799997] env[61986]: DEBUG nova.compute.provider_tree [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.858397] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159495, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.918408] env[61986]: DEBUG nova.network.neutron [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.007726] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.162383] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.267555] env[61986]: DEBUG nova.network.neutron [-] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.282396] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Successfully created port: 792a531c-fbc3-4df3-9ca9-f6a7f9d141db {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.302789] env[61986]: DEBUG nova.scheduler.client.report [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.361156] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159495, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.639979} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.361156] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6b586395-eb84-44f6-84c8-4a3ef3d64254/6b586395-eb84-44f6-84c8-4a3ef3d64254.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 609.361395] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 609.364211] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf9a8406-a354-45ed-9930-dce7c4cc564a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.372036] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 609.372036] env[61986]: value = "task-1159496" [ 609.372036] env[61986]: _type = "Task" [ 609.372036] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.386604] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.422500] env[61986]: DEBUG oslo_concurrency.lockutils [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] Releasing lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.422799] env[61986]: DEBUG nova.compute.manager [req-5716abf7-8f37-4bc0-b828-9ec9fc313bac req-e80e47a8-51b4-4b7c-af63-e06041434214 service nova] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Received event network-vif-deleted-a9a9f75e-f803-4732-b2e3-7aa9e414daf9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 609.688403] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.769662] env[61986]: INFO nova.compute.manager [-] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Took 1.06 seconds to deallocate network for instance. [ 609.772825] env[61986]: DEBUG nova.compute.claims [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 609.773204] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.814253] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.821s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.816486] env[61986]: ERROR nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Traceback (most recent call last): [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self.driver.spawn(context, instance, image_meta, [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] vm_ref = self.build_virtual_machine(instance, [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.816486] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] for vif in network_info: [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return self._sync_wrapper(fn, *args, **kwargs) [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self.wait() [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self[:] = self._gt.wait() [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return self._exit_event.wait() [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] result = hub.switch() [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.816822] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return self.greenlet.switch() [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] result = function(*args, **kwargs) [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] return func(*args, **kwargs) [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] raise e [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] nwinfo = self.network_api.allocate_for_instance( [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] created_port_ids = self._update_ports_for_instance( [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] with excutils.save_and_reraise_exception(): [ 609.817168] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] self.force_reraise() [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] raise self.value [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] updated_port = self._update_port( [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] _ensure_no_port_binding_failure(port) [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] raise exception.PortBindingFailed(port_id=port['id']) [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] nova.exception.PortBindingFailed: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. [ 609.817488] env[61986]: ERROR nova.compute.manager [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] [ 609.817806] env[61986]: DEBUG nova.compute.utils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.819039] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.657s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.821024] env[61986]: INFO nova.compute.claims [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.830165] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Build of instance 69e3c117-ee31-4d80-812b-da3c12fe19b9 was re-scheduled: Binding failed for port 1c0b8587-acc7-4d94-860c-e33e46a2e1ac, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.830165] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.830385] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquiring lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.830683] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Acquired lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.830881] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.883977] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059961} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.883977] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 609.886933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c75b6e5-c4c9-43ff-a4b3-5210cd3c1910 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.911412] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 6b586395-eb84-44f6-84c8-4a3ef3d64254/6b586395-eb84-44f6-84c8-4a3ef3d64254.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 609.911760] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec1f711e-3d24-4a9d-b68f-118c4ccaeeb4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.938739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "40b8df0a-36f1-4a4d-a975-f596901d1f98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.939168] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "40b8df0a-36f1-4a4d-a975-f596901d1f98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.940950] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 609.940950] env[61986]: value = "task-1159497" [ 609.940950] env[61986]: _type = "Task" [ 609.940950] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.954414] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159497, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.018405] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.053743] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.054458] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.054951] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.055307] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.055620] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.055913] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.056367] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.056675] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.059018] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.059018] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.059018] env[61986]: DEBUG nova.virt.hardware [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.059018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b9741e-3e0d-4198-8e58-5c88794e3563 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.069512] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a69e11-d3df-4fe5-9202-c71c4c5e9c6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.372897] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.442160] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.458448] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159497, 'name': ReconfigVM_Task, 'duration_secs': 0.287496} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.458804] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 6b586395-eb84-44f6-84c8-4a3ef3d64254/6b586395-eb84-44f6-84c8-4a3ef3d64254.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 610.460074] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91533696-b496-47d5-9977-e5d51f34e2c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.466247] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 610.466247] env[61986]: value = "task-1159498" [ 610.466247] env[61986]: _type = "Task" [ 610.466247] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.477492] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159498, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.597623] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.897351] env[61986]: DEBUG nova.compute.manager [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Received event network-changed-5b6cbc4a-c419-450e-afdf-da4251c8bb11 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.897351] env[61986]: DEBUG nova.compute.manager [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Refreshing instance network info cache due to event network-changed-5b6cbc4a-c419-450e-afdf-da4251c8bb11. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 610.897351] env[61986]: DEBUG oslo_concurrency.lockutils [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] Acquiring lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.897351] env[61986]: DEBUG oslo_concurrency.lockutils [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] Acquired lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.897351] env[61986]: DEBUG nova.network.neutron [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Refreshing network info cache for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.988230] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159498, 'name': Rename_Task, 'duration_secs': 0.399818} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.990074] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.990452] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 610.995770] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e1916cb-5475-4011-a717-a8a1e802eae3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.002176] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 611.002176] env[61986]: value = "task-1159499" [ 611.002176] env[61986]: _type = "Task" [ 611.002176] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.011487] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.061543] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41af247c-b3c6-40c9-8cda-89242b29ac9e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.069957] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e05d5d0-5fda-4697-92ca-81d3eb064b6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.103206] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Releasing lock "refresh_cache-69e3c117-ee31-4d80-812b-da3c12fe19b9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.103206] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.103437] env[61986]: DEBUG nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.103640] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.106233] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386d61ee-f3da-4e9e-9711-38660c6f6d3c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.114642] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85853e7d-4db3-48d2-8979-71aae96e43f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.129080] env[61986]: DEBUG nova.compute.provider_tree [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.175715] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.477040] env[61986]: DEBUG nova.network.neutron [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.518600] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159499, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.634522] env[61986]: DEBUG nova.scheduler.client.report [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.673517] env[61986]: DEBUG nova.network.neutron [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.678298] env[61986]: DEBUG nova.network.neutron [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.759766] env[61986]: ERROR nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 611.759766] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.759766] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.759766] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.759766] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.759766] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.759766] env[61986]: ERROR nova.compute.manager raise self.value [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.759766] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.759766] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.759766] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.761030] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.761030] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.761030] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 611.761030] env[61986]: ERROR nova.compute.manager [ 611.761030] env[61986]: Traceback (most recent call last): [ 611.761030] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.761030] env[61986]: listener.cb(fileno) [ 611.761030] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.761030] env[61986]: result = function(*args, **kwargs) [ 611.761030] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.761030] env[61986]: return func(*args, **kwargs) [ 611.761030] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.761030] env[61986]: raise e [ 611.761030] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.761030] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 611.761030] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.761030] env[61986]: created_port_ids = self._update_ports_for_instance( [ 611.761030] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.761030] env[61986]: with excutils.save_and_reraise_exception(): [ 611.761030] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.761030] env[61986]: self.force_reraise() [ 611.761030] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.761030] env[61986]: raise self.value [ 611.761030] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.761030] env[61986]: updated_port = self._update_port( [ 611.761030] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.761030] env[61986]: _ensure_no_port_binding_failure(port) [ 611.761030] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.761030] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.761994] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 611.761994] env[61986]: Removing descriptor: 15 [ 611.761994] env[61986]: ERROR nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Traceback (most recent call last): [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] yield resources [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self.driver.spawn(context, instance, image_meta, [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.761994] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] vm_ref = self.build_virtual_machine(instance, [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] for vif in network_info: [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return self._sync_wrapper(fn, *args, **kwargs) [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self.wait() [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self[:] = self._gt.wait() [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return self._exit_event.wait() [ 611.762397] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] result = hub.switch() [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return self.greenlet.switch() [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] result = function(*args, **kwargs) [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return func(*args, **kwargs) [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] raise e [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] nwinfo = self.network_api.allocate_for_instance( [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.762786] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] created_port_ids = self._update_ports_for_instance( [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] with excutils.save_and_reraise_exception(): [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self.force_reraise() [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] raise self.value [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] updated_port = self._update_port( [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] _ensure_no_port_binding_failure(port) [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.763222] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] raise exception.PortBindingFailed(port_id=port['id']) [ 611.764386] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 611.764386] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] [ 611.764386] env[61986]: INFO nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Terminating instance [ 611.764529] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.764778] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquired lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.764950] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 612.020509] env[61986]: DEBUG oslo_vmware.api [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159499, 'name': PowerOnVM_Task, 'duration_secs': 0.828662} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.021021] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 612.022369] env[61986]: INFO nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Took 9.65 seconds to spawn the instance on the hypervisor. [ 612.022369] env[61986]: DEBUG nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 612.024590] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58407404-33aa-42de-b3e4-14ff90a853c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.140032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.140946] env[61986]: DEBUG nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.144174] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.421s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.177126] env[61986]: DEBUG oslo_concurrency.lockutils [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] Releasing lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.177672] env[61986]: DEBUG nova.compute.manager [req-4a091083-747f-4f56-bb06-d292b63b4a75 req-08173346-0cc2-4321-98f8-1d734d085982 service nova] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Received event network-vif-deleted-5b6cbc4a-c419-450e-afdf-da4251c8bb11 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 612.182656] env[61986]: INFO nova.compute.manager [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] Took 1.08 seconds to deallocate network for instance. [ 612.389015] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.548768] env[61986]: INFO nova.compute.manager [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Took 17.11 seconds to build instance. [ 612.609166] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.654131] env[61986]: DEBUG nova.compute.utils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.659291] env[61986]: DEBUG nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 612.857990] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9614197a-8917-44dc-8c59-344e1ffb9171 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.866641] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69bb078-c701-4399-b2ee-a8406d77c70c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.902164] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455bf2bb-1153-43a8-a3d3-320643e8e791 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.910116] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef941ac-7f95-447b-a845-306298990ca0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.925874] env[61986]: DEBUG nova.compute.provider_tree [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.050606] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac7e4453-d950-4702-8279-38250e1737f2 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "6b586395-eb84-44f6-84c8-4a3ef3d64254" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.628s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.116302] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Releasing lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.117070] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.117070] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 613.117365] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c082174f-0ecd-4e6f-882a-9c76d830b1a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.127938] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfa87c9-681f-473e-adae-48709f620d1a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.153192] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 03489669-961b-4e5b-87b1-ef05bb59433b could not be found. [ 613.153670] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 613.153670] env[61986]: INFO nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.153771] env[61986]: DEBUG oslo.service.loopingcall [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.154074] env[61986]: DEBUG nova.compute.manager [-] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.154722] env[61986]: DEBUG nova.network.neutron [-] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.161301] env[61986]: DEBUG nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.200073] env[61986]: DEBUG nova.network.neutron [-] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.242586] env[61986]: INFO nova.scheduler.client.report [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Deleted allocations for instance 69e3c117-ee31-4d80-812b-da3c12fe19b9 [ 613.429704] env[61986]: DEBUG nova.scheduler.client.report [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.708119] env[61986]: DEBUG nova.network.neutron [-] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.753245] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a8211f5-4219-46be-8c10-fde23dca2a55 tempest-ServerDiagnosticsTest-1204128227 tempest-ServerDiagnosticsTest-1204128227-project-member] Lock "69e3c117-ee31-4d80-812b-da3c12fe19b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.071s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.753538] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "69e3c117-ee31-4d80-812b-da3c12fe19b9" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 16.467s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.753723] env[61986]: INFO nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 69e3c117-ee31-4d80-812b-da3c12fe19b9] During sync_power_state the instance has a pending task (networking). Skip. [ 613.753887] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "69e3c117-ee31-4d80-812b-da3c12fe19b9" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.768936] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquiring lock "e51ac29e-48f7-4589-ab67-c83a616cc591" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.769213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Lock "e51ac29e-48f7-4589-ab67-c83a616cc591" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.936363] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.792s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.937041] env[61986]: ERROR nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Traceback (most recent call last): [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self.driver.spawn(context, instance, image_meta, [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] vm_ref = self.build_virtual_machine(instance, [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.937041] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] for vif in network_info: [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return self._sync_wrapper(fn, *args, **kwargs) [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self.wait() [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self[:] = self._gt.wait() [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return self._exit_event.wait() [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] result = hub.switch() [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.937405] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return self.greenlet.switch() [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] result = function(*args, **kwargs) [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] return func(*args, **kwargs) [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] raise e [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] nwinfo = self.network_api.allocate_for_instance( [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] created_port_ids = self._update_ports_for_instance( [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] with excutils.save_and_reraise_exception(): [ 613.937727] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] self.force_reraise() [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] raise self.value [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] updated_port = self._update_port( [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] _ensure_no_port_binding_failure(port) [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] raise exception.PortBindingFailed(port_id=port['id']) [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] nova.exception.PortBindingFailed: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. [ 613.938046] env[61986]: ERROR nova.compute.manager [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] [ 613.938338] env[61986]: DEBUG nova.compute.utils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.938998] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.691s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.941601] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.941601] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 613.942107] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Build of instance f7f3a069-04da-4090-bbb1-1406799b5fdd was re-scheduled: Binding failed for port 6395ca6c-109f-419c-93a1-5552d949daa1, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.942601] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.942870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.943031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquired lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.943178] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.944176] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.952s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.946556] env[61986]: INFO nova.compute.claims [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.950297] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e044a3-2139-4bf0-afb2-203287ea9f6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.960050] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484932fa-f0a4-4837-a633-79c65220888b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.977781] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67a4303-f861-4149-9015-269a14b2b59c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.986298] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf416b0-1dce-45aa-bcb5-ebe859104c6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.024111] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181543MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 614.024111] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.172846] env[61986]: DEBUG nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.210601] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.211784] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.212549] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.212549] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.212549] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.212549] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.212801] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.212986] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.213185] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.213347] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.213736] env[61986]: DEBUG nova.virt.hardware [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.214159] env[61986]: INFO nova.compute.manager [-] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Took 1.06 seconds to deallocate network for instance. [ 614.214955] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405de349-317f-4c2d-9c8b-330249177f28 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.222147] env[61986]: DEBUG nova.compute.claims [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 614.222391] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.228251] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a022e0f8-5532-493e-9ac8-0e2f763a238f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.246043] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 614.254486] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Creating folder: Project (837bf83759304081adce0c4a677a31d3). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 614.255532] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b16c8285-42c1-46bc-8868-b3d2e9c7c212 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.267322] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Created folder: Project (837bf83759304081adce0c4a677a31d3) in parent group-v252271. [ 614.267544] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Creating folder: Instances. Parent ref: group-v252275. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 614.267800] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8f51e41-5776-4b0c-acc8-7d430ed8f15e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.272427] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.277012] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Created folder: Instances in parent group-v252275. [ 614.277877] env[61986]: DEBUG oslo.service.loopingcall [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.277877] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 614.278039] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49187fff-daed-42cc-8a81-58b3276c1c0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.297736] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 614.297736] env[61986]: value = "task-1159502" [ 614.297736] env[61986]: _type = "Task" [ 614.297736] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.306501] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159502, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.378875] env[61986]: DEBUG nova.compute.manager [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Received event network-changed-792a531c-fbc3-4df3-9ca9-f6a7f9d141db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.380633] env[61986]: DEBUG nova.compute.manager [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Refreshing instance network info cache due to event network-changed-792a531c-fbc3-4df3-9ca9-f6a7f9d141db. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 614.380633] env[61986]: DEBUG oslo_concurrency.lockutils [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] Acquiring lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.380633] env[61986]: DEBUG oslo_concurrency.lockutils [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] Acquired lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.380633] env[61986]: DEBUG nova.network.neutron [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Refreshing network info cache for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.512266] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.623389] env[61986]: DEBUG nova.compute.manager [None req-85bdc068-9796-4376-86cf-01e092b9d091 tempest-ServerDiagnosticsV248Test-2048354 tempest-ServerDiagnosticsV248Test-2048354-project-admin] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 614.630188] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f03814-5dcd-43e7-b0ac-4e70446d1fc1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.637583] env[61986]: INFO nova.compute.manager [None req-85bdc068-9796-4376-86cf-01e092b9d091 tempest-ServerDiagnosticsV248Test-2048354 tempest-ServerDiagnosticsV248Test-2048354-project-admin] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Retrieving diagnostics [ 614.638686] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d258cf2-3ba8-433a-809d-8d06249b2b06 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.716744] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.807184] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.811970] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159502, 'name': CreateVM_Task, 'duration_secs': 0.262317} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.812149] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 614.812549] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.812705] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.813139] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 614.813529] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a432785-619e-4bd8-98ff-c871f11028ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.818258] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 614.818258] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5247892a-e548-dfb4-a534-59f59d2b1e70" [ 614.818258] env[61986]: _type = "Task" [ 614.818258] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.826245] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5247892a-e548-dfb4-a534-59f59d2b1e70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.932759] env[61986]: DEBUG nova.network.neutron [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.190287] env[61986]: DEBUG nova.network.neutron [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.220057] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Releasing lock "refresh_cache-f7f3a069-04da-4090-bbb1-1406799b5fdd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.220057] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.220057] env[61986]: DEBUG nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.220057] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.228074] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708812fc-a3c3-4629-8c36-5a27ef8244e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.236985] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd627748-2ba6-4458-a728-37fbd43184ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.275638] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.281475] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd92fe3-4b69-40d4-9cb3-4f89eb1a7152 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.286146] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620ebd04-f346-4226-a209-9bddb5bfc92c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.303975] env[61986]: DEBUG nova.compute.provider_tree [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.330115] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5247892a-e548-dfb4-a534-59f59d2b1e70, 'name': SearchDatastore_Task, 'duration_secs': 0.008081} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.330592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.330684] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 615.332602] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.332602] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.332602] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 615.332602] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4da2dbcd-917e-4c79-b35f-22473864c536 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.340127] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 615.340313] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 615.342802] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08a65789-2e69-42f6-95de-a8801f86224f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.347941] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 615.347941] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c35df6-5bc5-254a-ad2a-3cc1d3eb6d1b" [ 615.347941] env[61986]: _type = "Task" [ 615.347941] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.356807] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c35df6-5bc5-254a-ad2a-3cc1d3eb6d1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.697554] env[61986]: DEBUG oslo_concurrency.lockutils [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] Releasing lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.697554] env[61986]: DEBUG nova.compute.manager [req-92d07ce0-339c-4d69-897a-23c0787f94a9 req-b230d657-a8ce-450e-948c-c655cebffacf service nova] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Received event network-vif-deleted-792a531c-fbc3-4df3-9ca9-f6a7f9d141db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.783211] env[61986]: DEBUG nova.network.neutron [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.808255] env[61986]: DEBUG nova.scheduler.client.report [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.864975] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c35df6-5bc5-254a-ad2a-3cc1d3eb6d1b, 'name': SearchDatastore_Task, 'duration_secs': 0.007909} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.866572] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7b2f8c9-77a6-4fa8-be7d-b307bc0ad664 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.874305] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 615.874305] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527224b7-c5dd-3885-58f8-101c8903ba23" [ 615.874305] env[61986]: _type = "Task" [ 615.874305] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.884360] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527224b7-c5dd-3885-58f8-101c8903ba23, 'name': SearchDatastore_Task, 'duration_secs': 0.008995} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.884752] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.885128] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 615.885541] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d502aebc-9717-4326-823c-f18b61c04ce8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.891757] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 615.891757] env[61986]: value = "task-1159503" [ 615.891757] env[61986]: _type = "Task" [ 615.891757] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.902044] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159503, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.289753] env[61986]: INFO nova.compute.manager [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] Took 1.07 seconds to deallocate network for instance. [ 616.313244] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.313923] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.318353] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.267s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.406986] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159503, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476883} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.407271] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 616.407597] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 616.407778] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c585422-6d31-4b76-8eef-777c639ee9e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.414198] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 616.414198] env[61986]: value = "task-1159504" [ 616.414198] env[61986]: _type = "Task" [ 616.414198] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.423179] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.825339] env[61986]: DEBUG nova.compute.utils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.832609] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.833810] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 616.932819] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159504, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069239} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.933094] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 616.933910] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d501ec26-6071-4b56-940c-3998320a24b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.975661] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 616.978672] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99c38446-4fe8-4bb5-9ac7-cadf56257238 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.998662] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 616.998662] env[61986]: value = "task-1159505" [ 616.998662] env[61986]: _type = "Task" [ 616.998662] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.007794] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159505, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.053281] env[61986]: DEBUG nova.policy [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e30d6afeb2148e6b4685ca968ecaeda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f02f9e0f1e804220898b725b08762d26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 617.087217] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27b6e48-beed-4624-9b20-d0abdf87626f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.094822] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cd2aa4-07c8-4c12-8842-0c1a549c64fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.137207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73371d8-a352-46f7-b0ba-807b8972990f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.148012] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa906029-c4b3-4a43-aa18-2e9f8d80fec8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.167438] env[61986]: DEBUG nova.compute.provider_tree [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.333767] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.343612] env[61986]: INFO nova.scheduler.client.report [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Deleted allocations for instance f7f3a069-04da-4090-bbb1-1406799b5fdd [ 617.514404] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159505, 'name': ReconfigVM_Task, 'duration_secs': 0.263483} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.514404] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 617.514404] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abbbf5ff-6459-4dd2-bb7c-38ae1b141acd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.521527] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 617.521527] env[61986]: value = "task-1159506" [ 617.521527] env[61986]: _type = "Task" [ 617.521527] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.534302] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159506, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.671524] env[61986]: DEBUG nova.scheduler.client.report [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.862118] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c10cc9e6-b64d-44c2-9f7a-c0e1bf69f46a tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "f7f3a069-04da-4090-bbb1-1406799b5fdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.256s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.862431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "f7f3a069-04da-4090-bbb1-1406799b5fdd" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.575s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.862830] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f8d02d1-5059-4e33-88e3-b675b9952739 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.871711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910b5f3e-ab3a-4ad8-b8ef-953c47ac0dff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.035265] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159506, 'name': Rename_Task, 'duration_secs': 0.226522} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.035819] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 618.037956] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d525664-58ac-45b6-9d22-9947fedd3bf8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.046490] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 618.046490] env[61986]: value = "task-1159507" [ 618.046490] env[61986]: _type = "Task" [ 618.046490] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.059875] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.179250] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.179933] env[61986]: ERROR nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Traceback (most recent call last): [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self.driver.spawn(context, instance, image_meta, [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] vm_ref = self.build_virtual_machine(instance, [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.179933] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] for vif in network_info: [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return self._sync_wrapper(fn, *args, **kwargs) [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self.wait() [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self[:] = self._gt.wait() [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return self._exit_event.wait() [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] result = hub.switch() [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.180286] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return self.greenlet.switch() [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] result = function(*args, **kwargs) [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] return func(*args, **kwargs) [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] raise e [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] nwinfo = self.network_api.allocate_for_instance( [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] created_port_ids = self._update_ports_for_instance( [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] with excutils.save_and_reraise_exception(): [ 618.180668] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] self.force_reraise() [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] raise self.value [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] updated_port = self._update_port( [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] _ensure_no_port_binding_failure(port) [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] raise exception.PortBindingFailed(port_id=port['id']) [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] nova.exception.PortBindingFailed: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. [ 618.180994] env[61986]: ERROR nova.compute.manager [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] [ 618.181283] env[61986]: DEBUG nova.compute.utils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.184038] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Build of instance 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a was re-scheduled: Binding failed for port a9a9f75e-f803-4732-b2e3-7aa9e414daf9, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.184038] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.184038] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.184038] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.184667] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.185953] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.498s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.187851] env[61986]: INFO nova.compute.claims [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.349734] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.390633] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.390979] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.390979] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.391952] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.391952] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.391952] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.392509] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.392509] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.392509] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.395559] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.395559] env[61986]: DEBUG nova.virt.hardware [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.396135] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512f5707-651a-49d8-98c0-ccd086520acf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.402876] env[61986]: INFO nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f7f3a069-04da-4090-bbb1-1406799b5fdd] During the sync_power process the instance has moved from host None to host cpu-1 [ 618.403172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "f7f3a069-04da-4090-bbb1-1406799b5fdd" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.541s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.408723] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a328cb-815c-4111-83e2-e71519cc499a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.556240] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159507, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.619832] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Successfully created port: 7d184833-c077-4bd1-9f2f-10e35e319c05 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.748672] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.061432] env[61986]: DEBUG oslo_vmware.api [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159507, 'name': PowerOnVM_Task, 'duration_secs': 0.72738} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.062990] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.064585] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 619.064585] env[61986]: INFO nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Took 4.89 seconds to spawn the instance on the hypervisor. [ 619.064691] env[61986]: DEBUG nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 619.067376] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fe0636-b9c6-4aab-8b5d-64d27432aa09 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.416437] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95fc85a-69d6-4501-9ab0-b866d7a4632a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.427370] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fddc8f4-4578-47d1-9601-235ef5c54222 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.468184] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58025487-ef94-41d6-987f-1518fe0e137c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.480246] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6818458b-7808-45b9-9e2a-24e19851b145 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.498882] env[61986]: DEBUG nova.compute.provider_tree [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.566190] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.566400] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.566574] env[61986]: DEBUG nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.566755] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.591389] env[61986]: INFO nova.compute.manager [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Took 17.45 seconds to build instance. [ 619.659422] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.002037] env[61986]: DEBUG nova.scheduler.client.report [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.093218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-052034f0-dc2e-4931-9175-592966bee0de tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "76b7db85-734f-40fe-8d58-2fba89bf205f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.964s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.170617] env[61986]: DEBUG nova.network.neutron [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.507759] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.508270] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 620.516662] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.742s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.674247] env[61986]: INFO nova.compute.manager [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a] Took 1.11 seconds to deallocate network for instance. [ 620.968303] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "5aa473df-1720-43f3-b3ca-5ab9ff565378" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.968540] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "5aa473df-1720-43f3-b3ca-5ab9ff565378" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.028821] env[61986]: DEBUG nova.compute.utils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.030479] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.030677] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.139638] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquiring lock "f5af7037-96fd-49f1-af80-a528737a3fa7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.139938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Lock "f5af7037-96fd-49f1-af80-a528737a3fa7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.232446] env[61986]: DEBUG nova.policy [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffed403e60f24e12a3d52d6de5e19b41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ec3507da1e14816b071bae4683c0c50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 621.250474] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abeed716-3e28-455a-ab05-4da66706607d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.261032] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68b5f02-e822-42fc-823c-113be28fde47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.297210] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a6c95f-b19f-4c62-b69b-f9460083bc9c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.306320] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11414cea-1210-4fe1-8ff0-ef1b2cd7425d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.321212] env[61986]: DEBUG nova.compute.provider_tree [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.472913] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.533483] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 621.643765] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.713160] env[61986]: INFO nova.scheduler.client.report [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted allocations for instance 1bd21abf-5421-4f51-a4e9-fecb2c98aa3a [ 621.825072] env[61986]: DEBUG nova.scheduler.client.report [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.998474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.175711] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.223027] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d16f276f-e871-4cc7-897f-3c305c7444fe tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "1bd21abf-5421-4f51-a4e9-fecb2c98aa3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.171s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.331232] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.815s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.332454] env[61986]: ERROR nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Traceback (most recent call last): [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self.driver.spawn(context, instance, image_meta, [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] vm_ref = self.build_virtual_machine(instance, [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.332454] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] for vif in network_info: [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return self._sync_wrapper(fn, *args, **kwargs) [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self.wait() [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self[:] = self._gt.wait() [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return self._exit_event.wait() [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] result = hub.switch() [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.332994] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return self.greenlet.switch() [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] result = function(*args, **kwargs) [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] return func(*args, **kwargs) [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] raise e [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] nwinfo = self.network_api.allocate_for_instance( [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] created_port_ids = self._update_ports_for_instance( [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] with excutils.save_and_reraise_exception(): [ 622.333631] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] self.force_reraise() [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] raise self.value [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] updated_port = self._update_port( [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] _ensure_no_port_binding_failure(port) [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] raise exception.PortBindingFailed(port_id=port['id']) [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] nova.exception.PortBindingFailed: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. [ 622.334183] env[61986]: ERROR nova.compute.manager [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] [ 622.336079] env[61986]: DEBUG nova.compute.utils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.336079] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.346s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.337165] env[61986]: INFO nova.compute.claims [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.342933] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Build of instance bb58f75b-97d0-4615-a014-3de3bf7dee2f was re-scheduled: Binding failed for port 5b6cbc4a-c419-450e-afdf-da4251c8bb11, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.343802] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.343802] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquiring lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.346131] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Acquired lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.346131] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.546081] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.575734] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.575955] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.575955] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.576449] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.576449] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.576449] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.576691] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.576742] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.577522] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.577522] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.577522] env[61986]: DEBUG nova.virt.hardware [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.579162] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc72cab3-04e0-4e72-9424-9542c1d3a291 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.586368] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c2fdb7-5727-4641-bc42-807b8b05946b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.909900] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.918737] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Successfully created port: 8a83e57b-4d3f-47cb-a0e4-93fef82d732c {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.338678] env[61986]: INFO nova.compute.manager [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Rebuilding instance [ 623.362109] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.411455] env[61986]: DEBUG nova.compute.manager [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 623.413032] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b557c61-0a4d-4586-a150-1411051febf4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.584550] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7559339b-e28a-4762-992c-d1c6418b0bbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.592941] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a2d150-d53c-4a6f-b798-e69e6bb77def {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.633032] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318431d2-9108-4810-95ba-08e73bb349b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.642091] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d20abc2-8b45-467b-aa49-4808dfbc8641 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.656139] env[61986]: DEBUG nova.compute.provider_tree [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.869144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Releasing lock "refresh_cache-bb58f75b-97d0-4615-a014-3de3bf7dee2f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.869402] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.869586] env[61986]: DEBUG nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.869755] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.931681] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 623.932225] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09c46a9a-ad17-4026-945f-3890bb64a0cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.935751] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.943422] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 623.943422] env[61986]: value = "task-1159508" [ 623.943422] env[61986]: _type = "Task" [ 623.943422] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.955844] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.989741] env[61986]: ERROR nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 623.989741] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.989741] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.989741] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.989741] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.989741] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.989741] env[61986]: ERROR nova.compute.manager raise self.value [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.989741] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.989741] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.989741] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.990251] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.990251] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.990251] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 623.990251] env[61986]: ERROR nova.compute.manager [ 623.990251] env[61986]: Traceback (most recent call last): [ 623.990251] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.990251] env[61986]: listener.cb(fileno) [ 623.990251] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.990251] env[61986]: result = function(*args, **kwargs) [ 623.990251] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.990251] env[61986]: return func(*args, **kwargs) [ 623.990251] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.990251] env[61986]: raise e [ 623.990251] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.990251] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 623.990251] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.990251] env[61986]: created_port_ids = self._update_ports_for_instance( [ 623.990251] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.990251] env[61986]: with excutils.save_and_reraise_exception(): [ 623.990251] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.990251] env[61986]: self.force_reraise() [ 623.990251] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.990251] env[61986]: raise self.value [ 623.990251] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.990251] env[61986]: updated_port = self._update_port( [ 623.990251] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.990251] env[61986]: _ensure_no_port_binding_failure(port) [ 623.990251] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.990251] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.991214] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 623.991214] env[61986]: Removing descriptor: 15 [ 623.991214] env[61986]: ERROR nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Traceback (most recent call last): [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] yield resources [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self.driver.spawn(context, instance, image_meta, [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.991214] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] vm_ref = self.build_virtual_machine(instance, [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] for vif in network_info: [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return self._sync_wrapper(fn, *args, **kwargs) [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self.wait() [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self[:] = self._gt.wait() [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return self._exit_event.wait() [ 623.991593] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] result = hub.switch() [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return self.greenlet.switch() [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] result = function(*args, **kwargs) [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return func(*args, **kwargs) [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] raise e [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] nwinfo = self.network_api.allocate_for_instance( [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.991952] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] created_port_ids = self._update_ports_for_instance( [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] with excutils.save_and_reraise_exception(): [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self.force_reraise() [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] raise self.value [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] updated_port = self._update_port( [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] _ensure_no_port_binding_failure(port) [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.992357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] raise exception.PortBindingFailed(port_id=port['id']) [ 623.992702] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 623.992702] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] [ 623.992702] env[61986]: INFO nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Terminating instance [ 623.996851] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.996851] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquired lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.996851] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.099160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "42698792-0555-45ae-a525-62265259bc17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.099160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "42698792-0555-45ae-a525-62265259bc17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.163571] env[61986]: DEBUG nova.scheduler.client.report [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.439229] env[61986]: DEBUG nova.network.neutron [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.443026] env[61986]: DEBUG nova.compute.manager [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Received event network-changed-7d184833-c077-4bd1-9f2f-10e35e319c05 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 624.443385] env[61986]: DEBUG nova.compute.manager [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Refreshing instance network info cache due to event network-changed-7d184833-c077-4bd1-9f2f-10e35e319c05. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 624.443878] env[61986]: DEBUG oslo_concurrency.lockutils [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] Acquiring lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.457134] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159508, 'name': PowerOffVM_Task, 'duration_secs': 0.1176} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.458714] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 624.460318] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 624.460659] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c9eac0-26cc-4134-8116-6788c4a70fe5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.469933] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 624.470194] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae325b04-2cb0-4f9a-8c51-b66986111903 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.501267] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 624.501475] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 624.501643] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleting the datastore file [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 624.502252] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-972fe05b-8211-4952-9b81-58801ccfc023 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.508097] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 624.508097] env[61986]: value = "task-1159510" [ 624.508097] env[61986]: _type = "Task" [ 624.508097] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.521422] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159510, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.584547] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.601550] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.668421] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.668987] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.671423] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.648s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.945873] env[61986]: INFO nova.compute.manager [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] [instance: bb58f75b-97d0-4615-a014-3de3bf7dee2f] Took 1.08 seconds to deallocate network for instance. [ 625.014515] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.021254] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159510, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089658} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.021490] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.021659] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 625.021819] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.133898] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.175755] env[61986]: DEBUG nova.compute.utils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.185138] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.185138] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 625.209269] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.209541] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.378418] env[61986]: DEBUG nova.policy [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e30d6afeb2148e6b4685ca968ecaeda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f02f9e0f1e804220898b725b08762d26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 625.516792] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Releasing lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.518240] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.518240] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 625.518514] env[61986]: DEBUG oslo_concurrency.lockutils [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] Acquired lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.518834] env[61986]: DEBUG nova.network.neutron [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Refreshing network info cache for port 7d184833-c077-4bd1-9f2f-10e35e319c05 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 625.520017] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dca178a4-6601-4a8c-b692-e97159545188 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.534499] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081ddfba-e54d-45f2-ae38-fb97e79ecfc5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.564442] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a048d234-fb8d-4904-a016-2d1c6bd6d103 could not be found. [ 625.564636] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.564804] env[61986]: INFO nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Took 0.05 seconds to destroy the instance on the hypervisor. [ 625.565056] env[61986]: DEBUG oslo.service.loopingcall [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.567148] env[61986]: DEBUG nova.compute.manager [-] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.567148] env[61986]: DEBUG nova.network.neutron [-] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.614126] env[61986]: DEBUG nova.network.neutron [-] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.693428] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.716082] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.733549] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6b586395-eb84-44f6-84c8-4a3ef3d64254 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 626.013840] env[61986]: INFO nova.scheduler.client.report [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Deleted allocations for instance bb58f75b-97d0-4615-a014-3de3bf7dee2f [ 626.082505] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.083068] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.084562] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.084562] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.084666] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.084742] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.085651] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.085651] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.085651] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.085651] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.085651] env[61986]: DEBUG nova.virt.hardware [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.086547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a78b28-f6d5-42d5-abaa-1e6dfdcf3c81 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.099468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228d9515-4d1c-4ebf-89a9-52eb53d39342 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.119564] env[61986]: DEBUG nova.network.neutron [-] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.122666] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.129981] env[61986]: DEBUG oslo.service.loopingcall [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.130850] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 626.131119] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7e25bf2-34f5-49a8-88ab-a4703786f17e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.154429] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.154429] env[61986]: value = "task-1159511" [ 626.154429] env[61986]: _type = "Task" [ 626.154429] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.166432] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159511, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.239043] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance bb58f75b-97d0-4615-a014-3de3bf7dee2f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 626.239043] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 03489669-961b-4e5b-87b1-ef05bb59433b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 626.239043] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 76b7db85-734f-40fe-8d58-2fba89bf205f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 626.239533] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a048d234-fb8d-4904-a016-2d1c6bd6d103 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 626.239533] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ab392d78-a11d-44d0-81d6-8afcb169e6e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 626.239533] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 40b8df0a-36f1-4a4d-a975-f596901d1f98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 626.263580] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.350446] env[61986]: DEBUG nova.network.neutron [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.530088] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d288f761-2001-4930-85e2-23a94e4dbffd tempest-ImagesOneServerTestJSON-1648383031 tempest-ImagesOneServerTestJSON-1648383031-project-member] Lock "bb58f75b-97d0-4615-a014-3de3bf7dee2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.619s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.537022] env[61986]: DEBUG nova.compute.manager [None req-b6b62dbd-0cde-4d75-aa28-293be2d31c3e tempest-ServerDiagnosticsV248Test-2048354 tempest-ServerDiagnosticsV248Test-2048354-project-admin] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 626.537022] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20804bd-28d2-4e5f-bd96-931da61b6840 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.544917] env[61986]: INFO nova.compute.manager [None req-b6b62dbd-0cde-4d75-aa28-293be2d31c3e tempest-ServerDiagnosticsV248Test-2048354 tempest-ServerDiagnosticsV248Test-2048354-project-admin] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Retrieving diagnostics [ 626.546087] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db1ae8c-494e-4cb5-aeb3-deb8357b0cdc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.631985] env[61986]: INFO nova.compute.manager [-] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Took 1.07 seconds to deallocate network for instance. [ 626.635319] env[61986]: DEBUG nova.compute.claims [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 626.635319] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.665265] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159511, 'name': CreateVM_Task, 'duration_secs': 0.301852} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.665615] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 626.666748] env[61986]: DEBUG oslo_vmware.service [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcb4f74-3931-4af1-ba23-ca8a8f2e31c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.675457] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.675757] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.676218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 626.676571] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84c9e362-6dd2-48ea-a528-086275ab767c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.682378] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 626.682378] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529f6926-dfe9-fe94-28f0-8bbebce1ebe3" [ 626.682378] env[61986]: _type = "Task" [ 626.682378] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.690555] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529f6926-dfe9-fe94-28f0-8bbebce1ebe3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.709909] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.745955] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance e51ac29e-48f7-4589-ab67-c83a616cc591 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 626.750655] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.750880] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.751038] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.751220] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.751359] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.751770] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.752017] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.752384] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.752596] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.752760] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.752927] env[61986]: DEBUG nova.virt.hardware [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.753801] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820fcaff-b410-4fb1-a9dd-4e0a62bbae67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.762111] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1527872-70f6-4315-a0b9-c5f8340966af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.789247] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquiring lock "49d08e2b-1be4-4adb-98a8-98efad8eb7ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.789247] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Lock "49d08e2b-1be4-4adb-98a8-98efad8eb7ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.934904] env[61986]: DEBUG nova.network.neutron [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.197874] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.200482] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.200482] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.200482] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.200482] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.200482] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-607c590c-4c31-4231-a204-d01be628732e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.218436] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.218667] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 627.222038] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6945bc4-26f1-4395-82e9-726ab3a14ca1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.231860] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-001db883-51c9-401c-94a9-a2b1c484209f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.240355] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Successfully created port: e4549b4f-a8bd-4f44-87d4-b5cf94934705 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.250965] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 5aa473df-1720-43f3-b3ca-5ab9ff565378 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 627.250965] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 627.250965] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520f0ec6-fed1-0b32-ccf3-b1059e024e56" [ 627.250965] env[61986]: _type = "Task" [ 627.250965] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.264137] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520f0ec6-fed1-0b32-ccf3-b1059e024e56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.291879] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.438241] env[61986]: DEBUG oslo_concurrency.lockutils [req-8f64b22d-dc68-44d7-8850-778439b15e3f req-a8fca514-d99c-420f-ab3c-8b87a724e662 service nova] Releasing lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.756106] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance f5af7037-96fd-49f1-af80-a528737a3fa7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 627.766765] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 627.766940] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Creating directory with path [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.767206] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b06994c1-5d50-44a7-83c4-2f6f4649a812 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.801291] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Created directory with path [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.801291] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Fetch image to [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 627.801291] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Downloading image file data 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk on the data store datastore1 {{(pid=61986) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 627.801291] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5bdf2d-abd1-4ae7-867a-c6aa70366d6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.821625] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4353b4b8-a1ba-453a-88df-f68a60fba2ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.830694] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bc44d0-c028-4daf-a310-e073a7bb439f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.837690] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.870937] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9f6c6f-99b2-47ef-a510-815a3a5d9cef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.878071] env[61986]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-17167bd5-aec3-4dad-8f49-ea65530e494a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.966488] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Downloading image file data 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to the data store datastore1 {{(pid=61986) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 628.065262] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 628.198967] env[61986]: ERROR nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 628.198967] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.198967] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.198967] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.198967] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.198967] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.198967] env[61986]: ERROR nova.compute.manager raise self.value [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.198967] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.198967] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.198967] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.199946] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.199946] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.199946] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 628.199946] env[61986]: ERROR nova.compute.manager [ 628.199946] env[61986]: Traceback (most recent call last): [ 628.199946] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.199946] env[61986]: listener.cb(fileno) [ 628.199946] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.199946] env[61986]: result = function(*args, **kwargs) [ 628.199946] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.199946] env[61986]: return func(*args, **kwargs) [ 628.199946] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.199946] env[61986]: raise e [ 628.199946] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.199946] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 628.199946] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.199946] env[61986]: created_port_ids = self._update_ports_for_instance( [ 628.199946] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.199946] env[61986]: with excutils.save_and_reraise_exception(): [ 628.199946] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.199946] env[61986]: self.force_reraise() [ 628.199946] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.199946] env[61986]: raise self.value [ 628.199946] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.199946] env[61986]: updated_port = self._update_port( [ 628.199946] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.199946] env[61986]: _ensure_no_port_binding_failure(port) [ 628.199946] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.199946] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.201195] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 628.201195] env[61986]: Removing descriptor: 14 [ 628.201195] env[61986]: ERROR nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Traceback (most recent call last): [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] yield resources [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self.driver.spawn(context, instance, image_meta, [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.201195] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] vm_ref = self.build_virtual_machine(instance, [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] for vif in network_info: [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return self._sync_wrapper(fn, *args, **kwargs) [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self.wait() [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self[:] = self._gt.wait() [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return self._exit_event.wait() [ 628.201642] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] result = hub.switch() [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return self.greenlet.switch() [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] result = function(*args, **kwargs) [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return func(*args, **kwargs) [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] raise e [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] nwinfo = self.network_api.allocate_for_instance( [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.202450] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] created_port_ids = self._update_ports_for_instance( [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] with excutils.save_and_reraise_exception(): [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self.force_reraise() [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] raise self.value [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] updated_port = self._update_port( [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] _ensure_no_port_binding_failure(port) [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.202803] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] raise exception.PortBindingFailed(port_id=port['id']) [ 628.203159] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 628.203159] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] [ 628.203159] env[61986]: INFO nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Terminating instance [ 628.203159] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquiring lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.203159] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquired lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.203159] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.262582] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 42698792-0555-45ae-a525-62265259bc17 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 628.388114] env[61986]: DEBUG nova.compute.manager [req-6199a757-5b0b-461d-b590-381586a4dbee req-970629f0-4a1b-43a2-99a8-918e063dada2 service nova] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Received event network-vif-deleted-7d184833-c077-4bd1-9f2f-10e35e319c05 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.752376] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.759910] env[61986]: DEBUG nova.compute.manager [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Received event network-changed-8a83e57b-4d3f-47cb-a0e4-93fef82d732c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.759910] env[61986]: DEBUG nova.compute.manager [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Refreshing instance network info cache due to event network-changed-8a83e57b-4d3f-47cb-a0e4-93fef82d732c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 628.759910] env[61986]: DEBUG oslo_concurrency.lockutils [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] Acquiring lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.768158] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 628.768158] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 628.768158] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 628.821256] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 628.822323] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 628.871724] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Downloaded image file data 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk on the data store datastore1 {{(pid=61986) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 628.878989] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 628.879277] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Copying Virtual Disk [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk to [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 628.882025] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-649540b1-1f72-4c71-9132-93b51612326b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.892746] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 628.892746] env[61986]: value = "task-1159512" [ 628.892746] env[61986]: _type = "Task" [ 628.892746] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.903061] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.093018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e711ef-df43-46f6-b05a-3c1d913b588c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.106865] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bc9b87-15ef-4112-be10-d1c25fae43e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.143821] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "6b586395-eb84-44f6-84c8-4a3ef3d64254" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.143821] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "6b586395-eb84-44f6-84c8-4a3ef3d64254" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.143821] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "6b586395-eb84-44f6-84c8-4a3ef3d64254-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.143821] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "6b586395-eb84-44f6-84c8-4a3ef3d64254-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.147684] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "6b586395-eb84-44f6-84c8-4a3ef3d64254-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.147684] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.147684] env[61986]: INFO nova.compute.manager [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Terminating instance [ 629.154281] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458d23f7-a603-4be5-a531-89a634bc0092 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.157263] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "refresh_cache-6b586395-eb84-44f6-84c8-4a3ef3d64254" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.157263] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquired lock "refresh_cache-6b586395-eb84-44f6-84c8-4a3ef3d64254" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.157263] env[61986]: DEBUG nova.network.neutron [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.164067] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3de518-4787-4b28-8836-86215f80ce80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.175397] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "10568cfc-3634-47d7-96d6-893c4fe01282" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.175397] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "10568cfc-3634-47d7-96d6-893c4fe01282" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.188940] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.405315] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159512, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.657271] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Releasing lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.657708] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.657935] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 629.658263] env[61986]: DEBUG oslo_concurrency.lockutils [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] Acquired lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.658437] env[61986]: DEBUG nova.network.neutron [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Refreshing network info cache for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 629.659734] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdb69a85-02ce-4469-a8cc-c4ccdfaf6ca6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.681649] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1275b9-1e47-4ba3-937d-61cb9d88a78b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.695543] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.709580] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab392d78-a11d-44d0-81d6-8afcb169e6e1 could not be found. [ 629.709799] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 629.710013] env[61986]: INFO nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 629.710286] env[61986]: DEBUG oslo.service.loopingcall [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.710446] env[61986]: DEBUG nova.compute.manager [-] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.710533] env[61986]: DEBUG nova.network.neutron [-] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.772822] env[61986]: DEBUG nova.network.neutron [-] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.822225] env[61986]: DEBUG nova.network.neutron [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.905459] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.736836} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.905729] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Copied Virtual Disk [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk to [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 629.905907] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleting the datastore file [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/tmp-sparse.vmdk {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 629.906164] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b239f23-cd66-42d8-b796-e36c8d75905c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.912014] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 629.912014] env[61986]: value = "task-1159513" [ 629.912014] env[61986]: _type = "Task" [ 629.912014] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.924060] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.030829] env[61986]: DEBUG nova.network.neutron [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.197141] env[61986]: DEBUG nova.network.neutron [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.201155] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 630.201360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.530s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.201844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.979s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.284605] env[61986]: DEBUG nova.network.neutron [-] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.423773] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024359} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.425363] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 630.425956] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Moving file from [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 to [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5. {{(pid=61986) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 630.426149] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-1b9b3cd7-b37d-4e24-8dea-051d25ef45bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.435988] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 630.435988] env[61986]: value = "task-1159514" [ 630.435988] env[61986]: _type = "Task" [ 630.435988] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.440559] env[61986]: ERROR nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 630.440559] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.440559] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.440559] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.440559] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.440559] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.440559] env[61986]: ERROR nova.compute.manager raise self.value [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.440559] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.440559] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.440559] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.442846] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.442846] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.442846] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 630.442846] env[61986]: ERROR nova.compute.manager [ 630.442846] env[61986]: Traceback (most recent call last): [ 630.442846] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.442846] env[61986]: listener.cb(fileno) [ 630.442846] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.442846] env[61986]: result = function(*args, **kwargs) [ 630.442846] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.442846] env[61986]: return func(*args, **kwargs) [ 630.442846] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.442846] env[61986]: raise e [ 630.442846] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.442846] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 630.442846] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.442846] env[61986]: created_port_ids = self._update_ports_for_instance( [ 630.442846] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.442846] env[61986]: with excutils.save_and_reraise_exception(): [ 630.442846] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.442846] env[61986]: self.force_reraise() [ 630.442846] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.442846] env[61986]: raise self.value [ 630.442846] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.442846] env[61986]: updated_port = self._update_port( [ 630.442846] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.442846] env[61986]: _ensure_no_port_binding_failure(port) [ 630.442846] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.442846] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.444589] env[61986]: nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 630.444589] env[61986]: Removing descriptor: 15 [ 630.446502] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "236ecde7-1250-4ce1-837c-8266b95dd513" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.446502] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "236ecde7-1250-4ce1-837c-8266b95dd513" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.447676] env[61986]: ERROR nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Traceback (most recent call last): [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] yield resources [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self.driver.spawn(context, instance, image_meta, [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] vm_ref = self.build_virtual_machine(instance, [ 630.447676] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] for vif in network_info: [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return self._sync_wrapper(fn, *args, **kwargs) [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self.wait() [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self[:] = self._gt.wait() [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return self._exit_event.wait() [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.449158] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] result = hub.switch() [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return self.greenlet.switch() [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] result = function(*args, **kwargs) [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return func(*args, **kwargs) [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] raise e [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] nwinfo = self.network_api.allocate_for_instance( [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] created_port_ids = self._update_ports_for_instance( [ 630.449584] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] with excutils.save_and_reraise_exception(): [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self.force_reraise() [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] raise self.value [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] updated_port = self._update_port( [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] _ensure_no_port_binding_failure(port) [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] raise exception.PortBindingFailed(port_id=port['id']) [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 630.449959] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] [ 630.450556] env[61986]: INFO nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Terminating instance [ 630.450591] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.450746] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquired lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.450923] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.456519] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159514, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.535249] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Releasing lock "refresh_cache-6b586395-eb84-44f6-84c8-4a3ef3d64254" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.535725] env[61986]: DEBUG nova.compute.manager [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.535908] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 630.536795] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa38eb80-b84e-4879-bc27-ab12625c2d55 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.545036] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 630.545270] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbc71d45-1443-4102-ae9a-32cae9af2723 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.555824] env[61986]: DEBUG oslo_vmware.api [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 630.555824] env[61986]: value = "task-1159515" [ 630.555824] env[61986]: _type = "Task" [ 630.555824] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.564928] env[61986]: DEBUG oslo_vmware.api [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.622922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquiring lock "f70adc74-3b24-434a-829a-13274bae40b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.623666] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Lock "f70adc74-3b24-434a-829a-13274bae40b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.667308] env[61986]: DEBUG nova.network.neutron [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.787600] env[61986]: INFO nova.compute.manager [-] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Took 1.08 seconds to deallocate network for instance. [ 630.791439] env[61986]: DEBUG nova.compute.claims [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 630.791612] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.946651] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159514, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029387} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.947049] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] File moved {{(pid=61986) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 630.948756] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Cleaning up location [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 630.948756] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleting the datastore file [datastore1] vmware_temp/712e68dd-0390-4e8d-ab8d-fca75293dcb0 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 630.948756] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-192c47b0-65c4-4584-9ca4-adf6b26f181a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.956261] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 630.956261] env[61986]: value = "task-1159516" [ 630.956261] env[61986]: _type = "Task" [ 630.956261] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.970220] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159516, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.987967] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.001427] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f33c405-6a58-4c17-89ca-d673fc025f2b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.008746] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3157da17-6ebc-4d71-b2e2-efb97d268a99 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.049271] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89db53ce-3ecb-433c-b002-146e1599d459 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.060869] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76885069-9576-427b-a18a-312ead29c697 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.070265] env[61986]: DEBUG oslo_vmware.api [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159515, 'name': PowerOffVM_Task, 'duration_secs': 0.125} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.072183] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 631.072365] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 631.072633] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96e24dcc-499e-4858-843a-db01a38d6d3f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.084682] env[61986]: DEBUG nova.compute.provider_tree [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.097514] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 631.100251] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 631.100251] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleting the datastore file [datastore2] 6b586395-eb84-44f6-84c8-4a3ef3d64254 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 631.100251] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9416ff25-4a5c-4304-a471-cc161957ff4a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.106969] env[61986]: DEBUG oslo_vmware.api [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for the task: (returnval){ [ 631.106969] env[61986]: value = "task-1159518" [ 631.106969] env[61986]: _type = "Task" [ 631.106969] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.115168] env[61986]: DEBUG oslo_vmware.api [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.171018] env[61986]: DEBUG oslo_concurrency.lockutils [req-db058670-2d90-4c5c-8814-f764fc58b953 req-08f48ad7-84f6-41ff-8eeb-fab548cc26ff service nova] Releasing lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.232854] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.467018] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159516, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027567} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.467643] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 631.469138] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4997cec4-d045-4505-aa0b-c4d00ecf48f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.475042] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 631.475042] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52665bbd-ab3e-c447-484d-d44b692b3a0c" [ 631.475042] env[61986]: _type = "Task" [ 631.475042] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.485107] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52665bbd-ab3e-c447-484d-d44b692b3a0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.589666] env[61986]: DEBUG nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.621864] env[61986]: DEBUG oslo_vmware.api [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Task: {'id': task-1159518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091012} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.622501] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 631.623092] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 631.623092] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 631.623092] env[61986]: INFO nova.compute.manager [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Took 1.09 seconds to destroy the instance on the hypervisor. [ 631.623327] env[61986]: DEBUG oslo.service.loopingcall [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.625025] env[61986]: DEBUG nova.compute.manager [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.625025] env[61986]: DEBUG nova.network.neutron [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.673146] env[61986]: DEBUG nova.network.neutron [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.735291] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Releasing lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.735765] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.735963] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 631.736283] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e65327f-36b1-493b-b2bc-3508c740d484 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.753363] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e51e976-afc8-4236-bc2d-d27c2c861a3f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.782735] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 40b8df0a-36f1-4a4d-a975-f596901d1f98 could not be found. [ 631.783530] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 631.783530] env[61986]: INFO nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Took 0.05 seconds to destroy the instance on the hypervisor. [ 631.784592] env[61986]: DEBUG oslo.service.loopingcall [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.784592] env[61986]: DEBUG nova.compute.manager [-] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.784592] env[61986]: DEBUG nova.network.neutron [-] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.835991] env[61986]: DEBUG nova.network.neutron [-] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.993018] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52665bbd-ab3e-c447-484d-d44b692b3a0c, 'name': SearchDatastore_Task, 'duration_secs': 0.010123} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.993338] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.993700] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 631.993914] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a1cd7fa-8159-4734-be90-afff62a5f7dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.002900] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 632.002900] env[61986]: value = "task-1159519" [ 632.002900] env[61986]: _type = "Task" [ 632.002900] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.016110] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159519, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.097933] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.098653] env[61986]: ERROR nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Traceback (most recent call last): [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self.driver.spawn(context, instance, image_meta, [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] vm_ref = self.build_virtual_machine(instance, [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.098653] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] for vif in network_info: [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return self._sync_wrapper(fn, *args, **kwargs) [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self.wait() [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self[:] = self._gt.wait() [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return self._exit_event.wait() [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] result = hub.switch() [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.099033] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return self.greenlet.switch() [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] result = function(*args, **kwargs) [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] return func(*args, **kwargs) [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] raise e [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] nwinfo = self.network_api.allocate_for_instance( [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] created_port_ids = self._update_ports_for_instance( [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] with excutils.save_and_reraise_exception(): [ 632.099453] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] self.force_reraise() [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] raise self.value [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] updated_port = self._update_port( [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] _ensure_no_port_binding_failure(port) [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] raise exception.PortBindingFailed(port_id=port['id']) [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] nova.exception.PortBindingFailed: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. [ 632.101361] env[61986]: ERROR nova.compute.manager [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] [ 632.102135] env[61986]: DEBUG nova.compute.utils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.103813] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.296s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.105724] env[61986]: INFO nova.compute.claims [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.113444] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Build of instance 03489669-961b-4e5b-87b1-ef05bb59433b was re-scheduled: Binding failed for port 792a531c-fbc3-4df3-9ca9-f6a7f9d141db, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 632.113444] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 632.113444] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.115124] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquired lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.115341] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 632.150113] env[61986]: DEBUG nova.compute.manager [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Received event network-changed-e4549b4f-a8bd-4f44-87d4-b5cf94934705 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 632.150113] env[61986]: DEBUG nova.compute.manager [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Refreshing instance network info cache due to event network-changed-e4549b4f-a8bd-4f44-87d4-b5cf94934705. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 632.150113] env[61986]: DEBUG oslo_concurrency.lockutils [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] Acquiring lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.150113] env[61986]: DEBUG oslo_concurrency.lockutils [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] Acquired lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.150113] env[61986]: DEBUG nova.network.neutron [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Refreshing network info cache for port e4549b4f-a8bd-4f44-87d4-b5cf94934705 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 632.181412] env[61986]: DEBUG nova.network.neutron [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.342653] env[61986]: DEBUG nova.network.neutron [-] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.404853] env[61986]: DEBUG nova.compute.manager [req-556fc7c9-6928-4ada-b3bf-f1b28374b526 req-9248ee77-419a-4977-b591-9d7f2cadb2ce service nova] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Received event network-vif-deleted-8a83e57b-4d3f-47cb-a0e4-93fef82d732c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 632.517022] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159519, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494561} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.517330] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 632.517539] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.517930] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb5a1332-4320-4804-ab2c-047f0869370e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.525640] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 632.525640] env[61986]: value = "task-1159520" [ 632.525640] env[61986]: _type = "Task" [ 632.525640] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.535803] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159520, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.685389] env[61986]: INFO nova.compute.manager [-] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Took 1.06 seconds to deallocate network for instance. [ 632.715180] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.789731] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquiring lock "35e41d19-8f1c-496d-9712-a1ac3c787c3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.790055] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Lock "35e41d19-8f1c-496d-9712-a1ac3c787c3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.790912] env[61986]: DEBUG nova.network.neutron [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.849615] env[61986]: INFO nova.compute.manager [-] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Took 1.06 seconds to deallocate network for instance. [ 632.851196] env[61986]: DEBUG nova.compute.claims [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 632.851589] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.000859] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquiring lock "4fa110e0-7235-4d62-a68b-c727b4a79908" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.001105] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Lock "4fa110e0-7235-4d62-a68b-c727b4a79908" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.038355] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159520, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140143} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.038625] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 633.039659] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a66263-6c2a-433b-8c92-05fd4a4d90ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.061315] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 633.061643] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f560a93-982a-4a94-8ea4-ff3950fdd00b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.084826] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 633.084826] env[61986]: value = "task-1159521" [ 633.084826] env[61986]: _type = "Task" [ 633.084826] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.095077] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159521, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.126549] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.172812] env[61986]: DEBUG nova.network.neutron [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.196698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.455139] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f8dec2-b137-4208-b165-83abae838599 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.467869] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8847c6-a298-4b20-8670-a6082bcf47d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.505705] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98358f67-88e2-4065-aadb-3711321c7502 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.514368] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd68451e-b41e-45d4-949b-8b6d86e74db0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.531636] env[61986]: DEBUG nova.compute.provider_tree [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.600052] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159521, 'name': ReconfigVM_Task, 'duration_secs': 0.497803} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.600459] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Reconfigured VM instance instance-00000007 to attach disk [datastore1] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.601085] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf7339e0-0ef3-4ac6-953d-3bc08cab75e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.611478] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 633.611478] env[61986]: value = "task-1159522" [ 633.611478] env[61986]: _type = "Task" [ 633.611478] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.622902] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159522, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.633808] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Releasing lock "refresh_cache-03489669-961b-4e5b-87b1-ef05bb59433b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.634173] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 633.634719] env[61986]: DEBUG nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.634905] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.669611] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.676803] env[61986]: DEBUG oslo_concurrency.lockutils [req-50523ddd-c3f1-48c9-918d-c20dcf553b4a req-87a7d280-78ec-445f-bc56-aa53fc7e4cdf service nova] Releasing lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.036783] env[61986]: DEBUG nova.scheduler.client.report [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.126084] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159522, 'name': Rename_Task, 'duration_secs': 0.136462} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.126678] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 634.126926] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-810cb2db-e6ce-4b78-8bd2-aad49801000a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.133940] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 634.133940] env[61986]: value = "task-1159523" [ 634.133940] env[61986]: _type = "Task" [ 634.133940] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.141969] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159523, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.176803] env[61986]: DEBUG nova.network.neutron [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.545281] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.545835] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.552444] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.554s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.553885] env[61986]: INFO nova.compute.claims [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.645658] env[61986]: DEBUG oslo_vmware.api [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159523, 'name': PowerOnVM_Task, 'duration_secs': 0.46951} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.646029] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 634.646343] env[61986]: DEBUG nova.compute.manager [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 634.648132] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f3a774-7762-49fd-8df7-42c75f4f8833 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.683136] env[61986]: INFO nova.compute.manager [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 03489669-961b-4e5b-87b1-ef05bb59433b] Took 1.05 seconds to deallocate network for instance. [ 635.063031] env[61986]: DEBUG nova.compute.utils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.065927] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 635.066319] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 635.096648] env[61986]: DEBUG nova.compute.manager [req-afdc06fe-823d-4aa7-888b-9af36877ae9e req-d60e744a-2fcb-4434-99a2-cd0bf85fe671 service nova] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Received event network-vif-deleted-e4549b4f-a8bd-4f44-87d4-b5cf94934705 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.123259] env[61986]: DEBUG nova.policy [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64d60e79d73c4777988a43d842e047c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f708a6fd32334ae4b7995f0de42f06a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 635.170737] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.516594] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquiring lock "b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.516964] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Lock "b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.567202] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.664927] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Successfully created port: 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.730941] env[61986]: INFO nova.scheduler.client.report [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Deleted allocations for instance 03489669-961b-4e5b-87b1-ef05bb59433b [ 635.925699] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f44aaaa-b3d0-4862-806c-963d29e64885 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.939096] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b162da1-5174-4683-8008-07c4af77eb38 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.971864] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff764e91-47ca-48e2-a5e1-8d3e1026baf1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.979766] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774af6b3-693e-4622-b414-4e93900d3e0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.995206] env[61986]: DEBUG nova.compute.provider_tree [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.240736] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8e02e94b-9eb1-42f8-aa66-68d8dc5ad5df tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "03489669-961b-4e5b-87b1-ef05bb59433b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.625s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.253914] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquiring lock "cbdf7e34-57f9-4456-8cc7-f09c15e0f965" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.254197] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Lock "cbdf7e34-57f9-4456-8cc7-f09c15e0f965" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.267791] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquiring lock "bf115dff-5440-41ba-9689-6337f63dedeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.267791] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Lock "bf115dff-5440-41ba-9689-6337f63dedeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.501986] env[61986]: DEBUG nova.scheduler.client.report [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.581838] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.618681] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:39:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='348885379',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1376419732',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.619459] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.619459] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.619459] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.620294] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.620501] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.620780] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.620952] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.621142] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.621307] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.621475] env[61986]: DEBUG nova.virt.hardware [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.622426] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9571e221-a02e-4e64-940d-e2044a4b45b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.632115] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9eaac8f-902b-4cd9-9ea1-a3412d57e628 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.744696] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.012739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.013351] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.016227] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.841s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.017642] env[61986]: INFO nova.compute.claims [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.266965] env[61986]: INFO nova.compute.manager [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Rebuilding instance [ 637.289943] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.348248] env[61986]: DEBUG nova.compute.manager [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 637.349205] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a144652-d1de-4f39-ba26-840be8a80316 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.376025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquiring lock "9f8b9aca-420f-4314-a466-0d6f3d191377" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.376025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Lock "9f8b9aca-420f-4314-a466-0d6f3d191377" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.407534] env[61986]: ERROR nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 637.407534] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.407534] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.407534] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.407534] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.407534] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.407534] env[61986]: ERROR nova.compute.manager raise self.value [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.407534] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.407534] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.407534] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.408092] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.408092] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.408092] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 637.408092] env[61986]: ERROR nova.compute.manager [ 637.408092] env[61986]: Traceback (most recent call last): [ 637.408092] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.408092] env[61986]: listener.cb(fileno) [ 637.408092] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.408092] env[61986]: result = function(*args, **kwargs) [ 637.408092] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.408092] env[61986]: return func(*args, **kwargs) [ 637.408092] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.408092] env[61986]: raise e [ 637.408092] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.408092] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 637.408092] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.408092] env[61986]: created_port_ids = self._update_ports_for_instance( [ 637.408092] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.408092] env[61986]: with excutils.save_and_reraise_exception(): [ 637.408092] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.408092] env[61986]: self.force_reraise() [ 637.408092] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.408092] env[61986]: raise self.value [ 637.408092] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.408092] env[61986]: updated_port = self._update_port( [ 637.408092] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.408092] env[61986]: _ensure_no_port_binding_failure(port) [ 637.408092] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.408092] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.408961] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 637.408961] env[61986]: Removing descriptor: 15 [ 637.408961] env[61986]: ERROR nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Traceback (most recent call last): [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] yield resources [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self.driver.spawn(context, instance, image_meta, [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.408961] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] vm_ref = self.build_virtual_machine(instance, [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] for vif in network_info: [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return self._sync_wrapper(fn, *args, **kwargs) [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self.wait() [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self[:] = self._gt.wait() [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return self._exit_event.wait() [ 637.409404] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] result = hub.switch() [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return self.greenlet.switch() [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] result = function(*args, **kwargs) [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return func(*args, **kwargs) [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] raise e [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] nwinfo = self.network_api.allocate_for_instance( [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.409811] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] created_port_ids = self._update_ports_for_instance( [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] with excutils.save_and_reraise_exception(): [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self.force_reraise() [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] raise self.value [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] updated_port = self._update_port( [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] _ensure_no_port_binding_failure(port) [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.410252] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] raise exception.PortBindingFailed(port_id=port['id']) [ 637.410630] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 637.410630] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] [ 637.410630] env[61986]: INFO nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Terminating instance [ 637.413371] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquiring lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.413371] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquired lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.413371] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.522443] env[61986]: DEBUG nova.compute.utils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.524808] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 637.524861] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 637.628946] env[61986]: DEBUG nova.policy [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3ca3baaab63400684218e32dc8b8bdf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ddd77efba5d94caf81e14827bbae6ed8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 637.866317] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 637.866317] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a5dab0c-6e26-4ba9-9899-47f4dead6ee6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.873674] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 637.873674] env[61986]: value = "task-1159524" [ 637.873674] env[61986]: _type = "Task" [ 637.873674] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.889044] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.959722] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.033737] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.209195] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.316810] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Successfully created port: 35badee2-138a-4abb-9304-06fde6862a96 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.394230] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159524, 'name': PowerOffVM_Task, 'duration_secs': 0.139092} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.394587] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 638.394587] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 638.395374] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a39407f-0d1d-4947-aa02-9ebd84552c7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.412285] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 638.412567] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquiring lock "a2c58c0a-785b-42a2-9960-0b3e4ff86d96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.412799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Lock "a2c58c0a-785b-42a2-9960-0b3e4ff86d96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.412935] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ade4c64b-666f-45aa-9ef0-a7bd3fc76b87 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.446906] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 638.447141] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 638.447664] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Deleting the datastore file [datastore1] 76b7db85-734f-40fe-8d58-2fba89bf205f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 638.447664] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-536ee711-d430-4057-b50d-67830d88154d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.457785] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 638.457785] env[61986]: value = "task-1159526" [ 638.457785] env[61986]: _type = "Task" [ 638.457785] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.467178] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39173763-10f5-4979-b099-60de32bbce0e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.475812] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.480174] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a529be4a-43ef-4c5e-a4fd-f8328440e7cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.513498] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07193ba0-c404-4c13-a656-4c11b0d3ca58 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.517051] env[61986]: DEBUG nova.compute.manager [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Received event network-changed-6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.517245] env[61986]: DEBUG nova.compute.manager [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Refreshing instance network info cache due to event network-changed-6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 638.517428] env[61986]: DEBUG oslo_concurrency.lockutils [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] Acquiring lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.523432] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e3d7ff-06b5-414d-a2f9-debdb347537d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.539013] env[61986]: DEBUG nova.compute.provider_tree [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.718222] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Releasing lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.719792] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.719926] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 638.720158] env[61986]: DEBUG oslo_concurrency.lockutils [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] Acquired lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.720802] env[61986]: DEBUG nova.network.neutron [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Refreshing network info cache for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 638.721708] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11eee77c-5dfa-4a34-a7fe-e07a41884792 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.735784] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a77199-f7c8-4019-9eab-dfbc3b9d9a25 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.757505] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e51ac29e-48f7-4589-ab67-c83a616cc591 could not be found. [ 638.757718] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 638.757962] env[61986]: INFO nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Took 0.04 seconds to destroy the instance on the hypervisor. [ 638.758150] env[61986]: DEBUG oslo.service.loopingcall [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.758356] env[61986]: DEBUG nova.compute.manager [-] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.758450] env[61986]: DEBUG nova.network.neutron [-] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.783847] env[61986]: DEBUG nova.network.neutron [-] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.971958] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122992} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.973017] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 638.973220] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 638.973418] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 639.047122] env[61986]: DEBUG nova.scheduler.client.report [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.052681] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.084410] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.084659] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.084865] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.084969] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.085125] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.085262] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.085676] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.085898] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.086221] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.086583] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.087017] env[61986]: DEBUG nova.virt.hardware [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.087657] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f00789-67f5-42de-ba2f-48d0b61ced6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.099457] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7988f9e-0afa-46ca-9510-e20cd7bd78a1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.251631] env[61986]: DEBUG nova.network.neutron [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.287092] env[61986]: DEBUG nova.network.neutron [-] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.358859] env[61986]: DEBUG nova.network.neutron [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.556822] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.556822] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 639.558545] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.425s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.561309] env[61986]: INFO nova.compute.claims [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.794555] env[61986]: INFO nova.compute.manager [-] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Took 1.04 seconds to deallocate network for instance. [ 639.797352] env[61986]: DEBUG nova.compute.claims [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 639.797580] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.862051] env[61986]: DEBUG oslo_concurrency.lockutils [req-adac7e83-0096-4b46-a25b-4d58bc2b00f7 req-32eea0f3-4e6b-4c4c-82ed-7173178ca0fd service nova] Releasing lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.023745] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 640.024086] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 640.025170] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.027698] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 640.027698] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.027698] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 640.027698] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 640.027698] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 640.028186] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 640.028186] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 640.028263] env[61986]: DEBUG nova.virt.hardware [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.033240] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d15116-68c2-4e9a-a92e-69239e4db5b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.046535] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ade86c0-9e73-4b26-b90a-6672923dc3c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.067056] env[61986]: DEBUG nova.compute.utils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.075645] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 640.077647] env[61986]: DEBUG oslo.service.loopingcall [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.077647] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.077647] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 640.082377] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 640.082377] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b3266a4-eb9d-4449-9b0d-78bc17e6d082 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.107568] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 640.107568] env[61986]: value = "task-1159527" [ 640.107568] env[61986]: _type = "Task" [ 640.107568] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.120299] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159527, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.124680] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.124680] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.169391] env[61986]: DEBUG nova.policy [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd8966765e9bb48a8acad2e890d329a31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '658839ed28d641c3be35ccb21e8bfe93', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 640.582613] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 640.624452] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159527, 'name': CreateVM_Task, 'duration_secs': 0.272807} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.624631] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 640.625592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.625592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.625592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 640.625803] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92156731-b15a-4888-9981-4bce91c4edc5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.638019] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 640.638019] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5266e179-005b-0095-e5de-532525bc9014" [ 640.638019] env[61986]: _type = "Task" [ 640.638019] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.646581] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5266e179-005b-0095-e5de-532525bc9014, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.754669] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "fec98558-ff3a-4d65-a650-42b6c3ab9e18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.754669] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "fec98558-ff3a-4d65-a650-42b6c3ab9e18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.809160] env[61986]: ERROR nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 640.809160] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.809160] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.809160] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.809160] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.809160] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.809160] env[61986]: ERROR nova.compute.manager raise self.value [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.809160] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.809160] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.809160] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.809715] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.809715] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.809715] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 640.809715] env[61986]: ERROR nova.compute.manager [ 640.809715] env[61986]: Traceback (most recent call last): [ 640.809715] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.809715] env[61986]: listener.cb(fileno) [ 640.809715] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.809715] env[61986]: result = function(*args, **kwargs) [ 640.809715] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.809715] env[61986]: return func(*args, **kwargs) [ 640.809715] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.809715] env[61986]: raise e [ 640.809715] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.809715] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 640.809715] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.809715] env[61986]: created_port_ids = self._update_ports_for_instance( [ 640.809715] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.809715] env[61986]: with excutils.save_and_reraise_exception(): [ 640.809715] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.809715] env[61986]: self.force_reraise() [ 640.809715] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.809715] env[61986]: raise self.value [ 640.809715] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.809715] env[61986]: updated_port = self._update_port( [ 640.809715] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.809715] env[61986]: _ensure_no_port_binding_failure(port) [ 640.809715] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.809715] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.810627] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 640.810627] env[61986]: Removing descriptor: 15 [ 640.810627] env[61986]: ERROR nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Traceback (most recent call last): [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] yield resources [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self.driver.spawn(context, instance, image_meta, [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.810627] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] vm_ref = self.build_virtual_machine(instance, [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] for vif in network_info: [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return self._sync_wrapper(fn, *args, **kwargs) [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self.wait() [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self[:] = self._gt.wait() [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return self._exit_event.wait() [ 640.811347] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] result = hub.switch() [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return self.greenlet.switch() [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] result = function(*args, **kwargs) [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return func(*args, **kwargs) [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] raise e [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] nwinfo = self.network_api.allocate_for_instance( [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.811767] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] created_port_ids = self._update_ports_for_instance( [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] with excutils.save_and_reraise_exception(): [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self.force_reraise() [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] raise self.value [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] updated_port = self._update_port( [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] _ensure_no_port_binding_failure(port) [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.812794] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] raise exception.PortBindingFailed(port_id=port['id']) [ 640.813214] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 640.813214] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] [ 640.813214] env[61986]: INFO nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Terminating instance [ 640.814853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.814853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.814992] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.953965] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d94f645-757f-48ea-b1e8-75b8b4075c64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.962516] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785d978e-1e0a-4e6c-accf-dfd859d45f15 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.997830] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f06f142-0475-45ee-94af-3ef0ab88ad46 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.004451] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17bfb13-b105-4c03-a777-38457198bd90 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.017320] env[61986]: DEBUG nova.compute.provider_tree [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.151331] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5266e179-005b-0095-e5de-532525bc9014, 'name': SearchDatastore_Task, 'duration_secs': 0.00997} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.153550] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.154053] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 641.154260] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.154870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.155387] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.155979] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25b7bda2-2aa4-47a7-bebb-c4ac9d4dcf08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.164858] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.165098] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 641.165787] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-018dead4-7c11-4060-b335-415d2231ad85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.172751] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 641.172751] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5284792f-935c-73df-8d56-61779262c81c" [ 641.172751] env[61986]: _type = "Task" [ 641.172751] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.182697] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5284792f-935c-73df-8d56-61779262c81c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.327351] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Successfully created port: 9686a128-bc3c-4094-9517-e0cfcd644c32 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.339836] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.399150] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.459229] env[61986]: DEBUG nova.compute.manager [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Received event network-vif-deleted-6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 641.459229] env[61986]: DEBUG nova.compute.manager [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Received event network-changed-35badee2-138a-4abb-9304-06fde6862a96 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 641.459229] env[61986]: DEBUG nova.compute.manager [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Refreshing instance network info cache due to event network-changed-35badee2-138a-4abb-9304-06fde6862a96. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 641.459397] env[61986]: DEBUG oslo_concurrency.lockutils [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] Acquiring lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.519745] env[61986]: DEBUG nova.scheduler.client.report [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.598915] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.635599] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.635599] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.635599] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.635768] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.635768] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.635768] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.639216] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.639471] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.639602] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.639778] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.640125] env[61986]: DEBUG nova.virt.hardware [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.640861] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e2b4ae-9eaa-419a-9e21-b5b33c83b702 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.650299] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452bde1f-d063-40ea-8ac1-102f7ad5923c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.684156] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5284792f-935c-73df-8d56-61779262c81c, 'name': SearchDatastore_Task, 'duration_secs': 0.008999} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.684950] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22802bf7-edde-408a-80d0-a8dd6e31550d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.690091] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 641.690091] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52216fc3-887d-c169-e0d3-1694f83aa3c6" [ 641.690091] env[61986]: _type = "Task" [ 641.690091] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.698204] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52216fc3-887d-c169-e0d3-1694f83aa3c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.904099] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.907050] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.907050] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 641.907050] env[61986]: DEBUG oslo_concurrency.lockutils [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] Acquired lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.907050] env[61986]: DEBUG nova.network.neutron [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Refreshing network info cache for port 35badee2-138a-4abb-9304-06fde6862a96 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 641.907050] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16f72713-c24d-4a76-87f3-7915efb91a25 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.918756] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fdb785-e453-48bb-b1a3-612646b17c30 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.943834] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5aa473df-1720-43f3-b3ca-5ab9ff565378 could not be found. [ 641.943982] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 641.944154] env[61986]: INFO nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Took 0.04 seconds to destroy the instance on the hypervisor. [ 641.944466] env[61986]: DEBUG oslo.service.loopingcall [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.944683] env[61986]: DEBUG nova.compute.manager [-] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.944778] env[61986]: DEBUG nova.network.neutron [-] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.961942] env[61986]: DEBUG nova.network.neutron [-] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.026592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.027330] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.031439] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.768s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.032933] env[61986]: INFO nova.compute.claims [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.202798] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52216fc3-887d-c169-e0d3-1694f83aa3c6, 'name': SearchDatastore_Task, 'duration_secs': 0.00995} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.204020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.204020] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 642.204020] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba6258b9-72fb-40bb-8888-470c4e162a26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.213708] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 642.213708] env[61986]: value = "task-1159528" [ 642.213708] env[61986]: _type = "Task" [ 642.213708] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.227982] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.400344] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "18f91633-5899-4f6d-addb-03c7ec39bc2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.400681] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "18f91633-5899-4f6d-addb-03c7ec39bc2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.464039] env[61986]: DEBUG nova.network.neutron [-] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.543321] env[61986]: DEBUG nova.compute.utils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 642.547403] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.547403] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 642.626412] env[61986]: DEBUG nova.network.neutron [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.724808] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495723} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.725917] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 642.725917] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.726354] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b84a810-6923-4cb5-9e65-d7336c5a3f84 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.735465] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 642.735465] env[61986]: value = "task-1159529" [ 642.735465] env[61986]: _type = "Task" [ 642.735465] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.741991] env[61986]: DEBUG nova.policy [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 642.750368] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.797280] env[61986]: DEBUG nova.network.neutron [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.968625] env[61986]: INFO nova.compute.manager [-] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Took 1.02 seconds to deallocate network for instance. [ 642.969384] env[61986]: DEBUG nova.compute.claims [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 642.970051] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.050351] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.255933] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063308} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.255933] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.255933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c2c47a-154f-4a74-a27d-9e177ef40b7d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.279219] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.282249] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a70e878-b755-4455-93dc-38e83d65a041 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.299694] env[61986]: DEBUG oslo_concurrency.lockutils [req-431f6ac2-c647-4c67-881c-bc336bf2dd83 req-774482d6-5a98-4e54-9fe3-455b9ddf68eb service nova] Releasing lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.302075] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 643.302075] env[61986]: value = "task-1159530" [ 643.302075] env[61986]: _type = "Task" [ 643.302075] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.316880] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159530, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.420391] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Successfully created port: 42873233-21b5-4ad3-a938-075026f9859e {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.516748] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856268ab-688f-4c7a-a76f-48fcfb507063 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.533390] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9433100e-eaab-47b9-a192-51fcc18c116e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.575246] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cde5fa-b3b4-49bc-9bb7-c4e313ef1167 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.583161] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc7cec2-9eeb-4af4-8dc2-88506d6caaec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.597533] env[61986]: DEBUG nova.compute.provider_tree [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.697322] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "213957b6-77a4-48fe-8bf5-f03bb67510c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.697618] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "213957b6-77a4-48fe-8bf5-f03bb67510c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.815943] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159530, 'name': ReconfigVM_Task, 'duration_secs': 0.301981} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.816555] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f/76b7db85-734f-40fe-8d58-2fba89bf205f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.817182] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c822770-ebb1-45e9-b5a4-c6c277b2583d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.823869] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 643.823869] env[61986]: value = "task-1159531" [ 643.823869] env[61986]: _type = "Task" [ 643.823869] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.833836] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159531, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.076111] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 644.103829] env[61986]: DEBUG nova.scheduler.client.report [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.111710] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.111931] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.112179] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.113588] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.113588] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.113588] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.113588] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.113588] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.114072] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.114072] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.114072] env[61986]: DEBUG nova.virt.hardware [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.114454] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef7616c-ed50-4b5c-a09e-666a94a0ea7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.123939] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb33ff2-dcd6-4b1e-b2ac-29e9cca988ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.245232] env[61986]: ERROR nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 644.245232] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.245232] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.245232] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.245232] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.245232] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.245232] env[61986]: ERROR nova.compute.manager raise self.value [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.245232] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.245232] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.245232] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.246721] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.246721] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.246721] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 644.246721] env[61986]: ERROR nova.compute.manager [ 644.246721] env[61986]: Traceback (most recent call last): [ 644.246721] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.246721] env[61986]: listener.cb(fileno) [ 644.246721] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.246721] env[61986]: result = function(*args, **kwargs) [ 644.246721] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.246721] env[61986]: return func(*args, **kwargs) [ 644.246721] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.246721] env[61986]: raise e [ 644.246721] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.246721] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 644.246721] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.246721] env[61986]: created_port_ids = self._update_ports_for_instance( [ 644.246721] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.246721] env[61986]: with excutils.save_and_reraise_exception(): [ 644.246721] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.246721] env[61986]: self.force_reraise() [ 644.246721] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.246721] env[61986]: raise self.value [ 644.246721] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.246721] env[61986]: updated_port = self._update_port( [ 644.246721] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.246721] env[61986]: _ensure_no_port_binding_failure(port) [ 644.246721] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.246721] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.247656] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 644.247656] env[61986]: Removing descriptor: 14 [ 644.247656] env[61986]: ERROR nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Traceback (most recent call last): [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] yield resources [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self.driver.spawn(context, instance, image_meta, [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.247656] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] vm_ref = self.build_virtual_machine(instance, [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] for vif in network_info: [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return self._sync_wrapper(fn, *args, **kwargs) [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self.wait() [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self[:] = self._gt.wait() [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return self._exit_event.wait() [ 644.248055] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] result = hub.switch() [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return self.greenlet.switch() [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] result = function(*args, **kwargs) [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return func(*args, **kwargs) [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] raise e [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] nwinfo = self.network_api.allocate_for_instance( [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.249549] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] created_port_ids = self._update_ports_for_instance( [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] with excutils.save_and_reraise_exception(): [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self.force_reraise() [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] raise self.value [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] updated_port = self._update_port( [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] _ensure_no_port_binding_failure(port) [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.250085] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] raise exception.PortBindingFailed(port_id=port['id']) [ 644.251219] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 644.251219] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] [ 644.251219] env[61986]: INFO nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Terminating instance [ 644.251219] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquiring lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.251219] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquired lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.251219] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.267624] env[61986]: DEBUG nova.compute.manager [req-791f3930-ed4a-4d98-aac3-65645bff58b3 req-0872a55b-05a7-434f-9873-e938c2885a3b service nova] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Received event network-vif-deleted-35badee2-138a-4abb-9304-06fde6862a96 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.284630] env[61986]: DEBUG nova.compute.manager [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Received event network-changed-9686a128-bc3c-4094-9517-e0cfcd644c32 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 644.284630] env[61986]: DEBUG nova.compute.manager [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Refreshing instance network info cache due to event network-changed-9686a128-bc3c-4094-9517-e0cfcd644c32. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 644.284792] env[61986]: DEBUG oslo_concurrency.lockutils [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] Acquiring lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.339066] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159531, 'name': Rename_Task, 'duration_secs': 0.134704} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.339066] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 644.339066] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18db961a-a4ae-4ab6-aa63-73232191cc68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.347138] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Waiting for the task: (returnval){ [ 644.347138] env[61986]: value = "task-1159532" [ 644.347138] env[61986]: _type = "Task" [ 644.347138] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.354516] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.620060] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.620595] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.623829] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.989s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.795887] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.857097] env[61986]: DEBUG oslo_vmware.api [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Task: {'id': task-1159532, 'name': PowerOnVM_Task, 'duration_secs': 0.496179} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.857097] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 644.857097] env[61986]: DEBUG nova.compute.manager [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 644.857097] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377d11d9-fb76-4e03-ab3b-de0ea53aacbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.972551] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.129985] env[61986]: DEBUG nova.compute.utils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.134978] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.134978] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 645.312427] env[61986]: DEBUG nova.policy [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48052f2800ad4caea81169b99fa55d1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'def2168f736745bba50861f9800a4c6b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 645.374445] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.476854] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Releasing lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.477707] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.477953] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 645.478212] env[61986]: DEBUG oslo_concurrency.lockutils [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] Acquired lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.481100] env[61986]: DEBUG nova.network.neutron [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Refreshing network info cache for port 9686a128-bc3c-4094-9517-e0cfcd644c32 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 645.481100] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6252eee-43d2-4f1a-8593-e54a1efd9aa4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.493338] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92deb2ba-e050-44be-943d-5266e31a7246 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.518672] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f5af7037-96fd-49f1-af80-a528737a3fa7 could not be found. [ 645.519672] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 645.519672] env[61986]: INFO nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 645.519672] env[61986]: DEBUG oslo.service.loopingcall [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.520533] env[61986]: DEBUG nova.compute.manager [-] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.525798] env[61986]: DEBUG nova.network.neutron [-] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.558248] env[61986]: DEBUG nova.network.neutron [-] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.631654] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03019e73-bcf5-414a-aa45-6a6dccc90dfb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.636710] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.644454] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b7485c-2fbc-4300-87fc-10ab18bb5dd0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.683810] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0771e8ac-4546-46ef-b645-fa69bb31b253 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.691705] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6edb80-07fb-4f2e-bad8-81f1e1990c91 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.712437] env[61986]: DEBUG nova.compute.provider_tree [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.888370] env[61986]: ERROR nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 645.888370] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.888370] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.888370] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.888370] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.888370] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.888370] env[61986]: ERROR nova.compute.manager raise self.value [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.888370] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.888370] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.888370] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.889194] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.889194] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.889194] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 645.889194] env[61986]: ERROR nova.compute.manager [ 645.889194] env[61986]: Traceback (most recent call last): [ 645.889194] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.889194] env[61986]: listener.cb(fileno) [ 645.889194] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.889194] env[61986]: result = function(*args, **kwargs) [ 645.889194] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.889194] env[61986]: return func(*args, **kwargs) [ 645.889194] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.889194] env[61986]: raise e [ 645.889194] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.889194] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 645.889194] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.889194] env[61986]: created_port_ids = self._update_ports_for_instance( [ 645.889194] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.889194] env[61986]: with excutils.save_and_reraise_exception(): [ 645.889194] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.889194] env[61986]: self.force_reraise() [ 645.889194] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.889194] env[61986]: raise self.value [ 645.889194] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.889194] env[61986]: updated_port = self._update_port( [ 645.889194] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.889194] env[61986]: _ensure_no_port_binding_failure(port) [ 645.889194] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.889194] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.890230] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 645.890230] env[61986]: Removing descriptor: 17 [ 645.890230] env[61986]: ERROR nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] Traceback (most recent call last): [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] yield resources [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self.driver.spawn(context, instance, image_meta, [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.890230] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] vm_ref = self.build_virtual_machine(instance, [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] for vif in network_info: [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return self._sync_wrapper(fn, *args, **kwargs) [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self.wait() [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self[:] = self._gt.wait() [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return self._exit_event.wait() [ 645.891932] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] result = hub.switch() [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return self.greenlet.switch() [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] result = function(*args, **kwargs) [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return func(*args, **kwargs) [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] raise e [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] nwinfo = self.network_api.allocate_for_instance( [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.892990] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] created_port_ids = self._update_ports_for_instance( [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] with excutils.save_and_reraise_exception(): [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self.force_reraise() [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] raise self.value [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] updated_port = self._update_port( [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] _ensure_no_port_binding_failure(port) [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.893830] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] raise exception.PortBindingFailed(port_id=port['id']) [ 645.895265] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 645.895265] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] [ 645.895265] env[61986]: INFO nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Terminating instance [ 645.895265] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.895265] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.895265] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.027126] env[61986]: DEBUG nova.network.neutron [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.063815] env[61986]: DEBUG nova.network.neutron [-] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.214454] env[61986]: DEBUG nova.scheduler.client.report [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.280643] env[61986]: DEBUG nova.network.neutron [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.296386] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Successfully created port: b5d913fb-250b-4705-b9f6-35103c57c660 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.432391] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.569443] env[61986]: INFO nova.compute.manager [-] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Took 1.05 seconds to deallocate network for instance. [ 646.572068] env[61986]: DEBUG nova.compute.claims [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 646.572270] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.651288] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.661646] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.695382] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.695597] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.695748] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.695926] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.696736] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.697055] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.697247] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.700151] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.700151] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.700151] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.700151] env[61986]: DEBUG nova.virt.hardware [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.700151] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21415160-cc3c-4047-906a-3947cc362c75 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.712060] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ad8cab-18c2-40cd-a2ab-376c6678293e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.731332] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.107s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.731971] env[61986]: ERROR nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Traceback (most recent call last): [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self.driver.spawn(context, instance, image_meta, [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] vm_ref = self.build_virtual_machine(instance, [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.731971] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] for vif in network_info: [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return self._sync_wrapper(fn, *args, **kwargs) [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self.wait() [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self[:] = self._gt.wait() [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return self._exit_event.wait() [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] result = hub.switch() [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.732357] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return self.greenlet.switch() [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] result = function(*args, **kwargs) [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] return func(*args, **kwargs) [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] raise e [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] nwinfo = self.network_api.allocate_for_instance( [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] created_port_ids = self._update_ports_for_instance( [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] with excutils.save_and_reraise_exception(): [ 646.732768] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] self.force_reraise() [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] raise self.value [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] updated_port = self._update_port( [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] _ensure_no_port_binding_failure(port) [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] raise exception.PortBindingFailed(port_id=port['id']) [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] nova.exception.PortBindingFailed: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. [ 646.733206] env[61986]: ERROR nova.compute.manager [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] [ 646.734390] env[61986]: DEBUG nova.compute.utils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.738201] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Build of instance a048d234-fb8d-4904-a016-2d1c6bd6d103 was re-scheduled: Binding failed for port 7d184833-c077-4bd1-9f2f-10e35e319c05, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.738201] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 646.738201] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.738201] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquired lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.738635] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.738635] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.901s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.739557] env[61986]: INFO nova.compute.claims [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.783667] env[61986]: DEBUG oslo_concurrency.lockutils [req-7738b1b6-0176-42d3-b943-2faa979aaf69 req-7cd99b95-b09a-4013-a28b-3e9c29d62095 service nova] Releasing lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.170576] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.170576] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 647.170576] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 647.172524] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-537f09a9-3066-4b39-bef2-201085de3154 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.177288] env[61986]: DEBUG nova.compute.manager [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] [instance: 42698792-0555-45ae-a525-62265259bc17] Received event network-changed-42873233-21b5-4ad3-a938-075026f9859e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.177447] env[61986]: DEBUG nova.compute.manager [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] [instance: 42698792-0555-45ae-a525-62265259bc17] Refreshing instance network info cache due to event network-changed-42873233-21b5-4ad3-a938-075026f9859e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 647.181017] env[61986]: DEBUG oslo_concurrency.lockutils [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] Acquiring lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.181374] env[61986]: DEBUG oslo_concurrency.lockutils [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] Acquired lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.181594] env[61986]: DEBUG nova.network.neutron [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] [instance: 42698792-0555-45ae-a525-62265259bc17] Refreshing network info cache for port 42873233-21b5-4ad3-a938-075026f9859e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.195858] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41890c45-5257-40b9-b51b-79c47ce4ff16 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.211052] env[61986]: DEBUG nova.compute.manager [req-f24c3812-23b7-415c-bbd7-a4d321320e36 req-2b8e1466-f43e-42a4-a05a-641446ee6aaa service nova] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Received event network-vif-deleted-9686a128-bc3c-4094-9517-e0cfcd644c32 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.226161] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 42698792-0555-45ae-a525-62265259bc17 could not be found. [ 647.226384] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 647.226584] env[61986]: INFO nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Took 0.06 seconds to destroy the instance on the hypervisor. [ 647.226862] env[61986]: DEBUG oslo.service.loopingcall [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.227178] env[61986]: DEBUG nova.compute.manager [-] [instance: 42698792-0555-45ae-a525-62265259bc17] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.227178] env[61986]: DEBUG nova.network.neutron [-] [instance: 42698792-0555-45ae-a525-62265259bc17] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.334436] env[61986]: DEBUG nova.network.neutron [-] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.340447] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.448032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "76b7db85-734f-40fe-8d58-2fba89bf205f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.450697] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "76b7db85-734f-40fe-8d58-2fba89bf205f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.450697] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "76b7db85-734f-40fe-8d58-2fba89bf205f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.450697] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "76b7db85-734f-40fe-8d58-2fba89bf205f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.450697] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "76b7db85-734f-40fe-8d58-2fba89bf205f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.453571] env[61986]: INFO nova.compute.manager [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Terminating instance [ 647.456340] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "refresh_cache-76b7db85-734f-40fe-8d58-2fba89bf205f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.456340] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquired lock "refresh_cache-76b7db85-734f-40fe-8d58-2fba89bf205f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.456340] env[61986]: DEBUG nova.network.neutron [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 647.623467] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.716028] env[61986]: DEBUG nova.network.neutron [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.837185] env[61986]: DEBUG nova.network.neutron [-] [instance: 42698792-0555-45ae-a525-62265259bc17] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.994461] env[61986]: DEBUG nova.network.neutron [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.114306] env[61986]: DEBUG nova.network.neutron [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] [instance: 42698792-0555-45ae-a525-62265259bc17] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.127399] env[61986]: DEBUG nova.network.neutron [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.129514] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Releasing lock "refresh_cache-a048d234-fb8d-4904-a016-2d1c6bd6d103" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.129727] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 648.131056] env[61986]: DEBUG nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.131056] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.153641] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.222529] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7c57f2-8f4d-4edf-ae7a-c36efd5a3280 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.231487] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f96ba6-e9d3-4a6a-924e-6cb82075eddd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.271399] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63e7e7e-d4a7-48b0-b19c-3e090679cf81 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.279543] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add5ae7d-6294-44fa-aced-5972ec06cdc2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.292939] env[61986]: DEBUG nova.compute.provider_tree [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.342747] env[61986]: INFO nova.compute.manager [-] [instance: 42698792-0555-45ae-a525-62265259bc17] Took 1.11 seconds to deallocate network for instance. [ 648.344676] env[61986]: DEBUG nova.compute.claims [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 648.344676] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.620162] env[61986]: DEBUG oslo_concurrency.lockutils [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] Releasing lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.620883] env[61986]: DEBUG nova.compute.manager [req-007991a1-ec6c-4eeb-8987-ad9e04d049d4 req-475ca134-000f-4d25-ab55-f6182cc5a2ce service nova] [instance: 42698792-0555-45ae-a525-62265259bc17] Received event network-vif-deleted-42873233-21b5-4ad3-a938-075026f9859e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 648.634231] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Releasing lock "refresh_cache-76b7db85-734f-40fe-8d58-2fba89bf205f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.634640] env[61986]: DEBUG nova.compute.manager [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.634833] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 648.635897] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6448f999-6595-4887-be13-d894d6def379 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.643939] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 648.643939] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26b28e7e-8799-491e-aba9-f421731dc708 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.650192] env[61986]: DEBUG oslo_vmware.api [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 648.650192] env[61986]: value = "task-1159533" [ 648.650192] env[61986]: _type = "Task" [ 648.650192] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.657989] env[61986]: DEBUG nova.network.neutron [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.660197] env[61986]: DEBUG oslo_vmware.api [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.684929] env[61986]: ERROR nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 648.684929] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.684929] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.684929] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.684929] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.684929] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.684929] env[61986]: ERROR nova.compute.manager raise self.value [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.684929] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.684929] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.684929] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.685528] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.685528] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.685528] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 648.685528] env[61986]: ERROR nova.compute.manager [ 648.685528] env[61986]: Traceback (most recent call last): [ 648.685528] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.685528] env[61986]: listener.cb(fileno) [ 648.685528] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.685528] env[61986]: result = function(*args, **kwargs) [ 648.685528] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.685528] env[61986]: return func(*args, **kwargs) [ 648.685528] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.685528] env[61986]: raise e [ 648.685528] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.685528] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 648.685528] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.685528] env[61986]: created_port_ids = self._update_ports_for_instance( [ 648.685528] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.685528] env[61986]: with excutils.save_and_reraise_exception(): [ 648.685528] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.685528] env[61986]: self.force_reraise() [ 648.685528] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.685528] env[61986]: raise self.value [ 648.685528] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.685528] env[61986]: updated_port = self._update_port( [ 648.685528] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.685528] env[61986]: _ensure_no_port_binding_failure(port) [ 648.685528] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.685528] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.686532] env[61986]: nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 648.686532] env[61986]: Removing descriptor: 14 [ 648.686653] env[61986]: ERROR nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Traceback (most recent call last): [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] yield resources [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self.driver.spawn(context, instance, image_meta, [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] vm_ref = self.build_virtual_machine(instance, [ 648.686653] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] for vif in network_info: [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return self._sync_wrapper(fn, *args, **kwargs) [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self.wait() [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self[:] = self._gt.wait() [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return self._exit_event.wait() [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.687034] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] result = hub.switch() [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return self.greenlet.switch() [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] result = function(*args, **kwargs) [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return func(*args, **kwargs) [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] raise e [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] nwinfo = self.network_api.allocate_for_instance( [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] created_port_ids = self._update_ports_for_instance( [ 648.687461] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] with excutils.save_and_reraise_exception(): [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self.force_reraise() [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] raise self.value [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] updated_port = self._update_port( [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] _ensure_no_port_binding_failure(port) [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] raise exception.PortBindingFailed(port_id=port['id']) [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 648.688452] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] [ 648.688979] env[61986]: INFO nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Terminating instance [ 648.694523] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.694523] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquired lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.698174] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.802319] env[61986]: DEBUG nova.scheduler.client.report [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.164338] env[61986]: DEBUG oslo_vmware.api [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159533, 'name': PowerOffVM_Task, 'duration_secs': 0.135047} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.164338] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 649.164338] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 649.164338] env[61986]: INFO nova.compute.manager [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: a048d234-fb8d-4904-a016-2d1c6bd6d103] Took 1.03 seconds to deallocate network for instance. [ 649.166846] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae62e471-3b35-4ef0-a5f8-e5231c87374e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.195695] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 649.195695] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 649.195695] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleting the datastore file [datastore2] 76b7db85-734f-40fe-8d58-2fba89bf205f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.196407] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a4eba6e-dc7f-4194-93fa-d29c8edfece8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.205214] env[61986]: DEBUG oslo_vmware.api [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for the task: (returnval){ [ 649.205214] env[61986]: value = "task-1159535" [ 649.205214] env[61986]: _type = "Task" [ 649.205214] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.217524] env[61986]: DEBUG oslo_vmware.api [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159535, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.238304] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.309096] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.309096] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.315027] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.316105] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.524s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.431955] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquiring lock "dd239b21-fdb2-48fb-acf4-18242a98cd1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.431955] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Lock "dd239b21-fdb2-48fb-acf4-18242a98cd1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.718302] env[61986]: DEBUG oslo_vmware.api [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Task: {'id': task-1159535, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096313} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.718594] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 649.718776] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 649.718976] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 649.719224] env[61986]: INFO nova.compute.manager [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Took 1.08 seconds to destroy the instance on the hypervisor. [ 649.719460] env[61986]: DEBUG oslo.service.loopingcall [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.722571] env[61986]: DEBUG nova.compute.manager [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.722571] env[61986]: DEBUG nova.network.neutron [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.746490] env[61986]: DEBUG nova.compute.manager [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Received event network-changed-b5d913fb-250b-4705-b9f6-35103c57c660 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.746684] env[61986]: DEBUG nova.compute.manager [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Refreshing instance network info cache due to event network-changed-b5d913fb-250b-4705-b9f6-35103c57c660. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 649.746868] env[61986]: DEBUG oslo_concurrency.lockutils [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] Acquiring lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.749387] env[61986]: DEBUG nova.network.neutron [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.822735] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Releasing lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.822735] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 649.822735] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 649.822735] env[61986]: DEBUG nova.compute.utils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.824329] env[61986]: DEBUG oslo_concurrency.lockutils [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] Acquired lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.824733] env[61986]: DEBUG nova.network.neutron [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Refreshing network info cache for port b5d913fb-250b-4705-b9f6-35103c57c660 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 649.825809] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f456ece2-760e-4ba1-ad89-2a7afed6f973 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.834838] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.834838] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 649.852338] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f69464-0926-4309-9b3a-83540284fccd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.885370] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a could not be found. [ 649.885370] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 649.885370] env[61986]: INFO nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Took 0.06 seconds to destroy the instance on the hypervisor. [ 649.885370] env[61986]: DEBUG oslo.service.loopingcall [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.885370] env[61986]: DEBUG nova.compute.manager [-] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.885370] env[61986]: DEBUG nova.network.neutron [-] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.910470] env[61986]: DEBUG nova.policy [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b01f5abf0f140358ba5cdeb4759dc2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eae4786dfa404dc29feda125a4665ea9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 649.923178] env[61986]: DEBUG nova.network.neutron [-] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.209802] env[61986]: INFO nova.scheduler.client.report [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Deleted allocations for instance a048d234-fb8d-4904-a016-2d1c6bd6d103 [ 650.253502] env[61986]: DEBUG nova.network.neutron [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.341597] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.347346] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdd7fb9-d592-49c6-a6d1-cd297779a781 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.356214] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58231b81-86e3-4037-aa7e-753b7148e035 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.393633] env[61986]: DEBUG nova.network.neutron [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.397224] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9abe130-e79a-4767-b878-fe1637f627e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.406199] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafbbfe3-7e50-4565-88b2-23b441f135d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.422441] env[61986]: DEBUG nova.compute.provider_tree [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.423625] env[61986]: DEBUG nova.network.neutron [-] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.525899] env[61986]: DEBUG nova.network.neutron [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.655604] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Successfully created port: d5b88187-18fc-4690-9ab8-e8c53d1a6a95 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.734595] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a06abd1c-6222-402d-a2de-ad73d3fff5a0 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "a048d234-fb8d-4904-a016-2d1c6bd6d103" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.773s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.756385] env[61986]: INFO nova.compute.manager [-] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Took 1.03 seconds to deallocate network for instance. [ 650.926364] env[61986]: DEBUG nova.scheduler.client.report [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.930976] env[61986]: INFO nova.compute.manager [-] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Took 1.05 seconds to deallocate network for instance. [ 650.934762] env[61986]: DEBUG nova.compute.claims [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 650.935104] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.034554] env[61986]: DEBUG oslo_concurrency.lockutils [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] Releasing lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.034826] env[61986]: DEBUG nova.compute.manager [req-cd2dbeb6-d665-44d0-8635-31aa760926aa req-268be751-ee72-42ab-afef-fadcc2bd3b51 service nova] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Received event network-vif-deleted-b5d913fb-250b-4705-b9f6-35103c57c660 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 651.236845] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.270731] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.359075] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.393363] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.393363] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.393957] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.394019] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.394185] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.394437] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.394915] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.395160] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.395397] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.395661] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.396100] env[61986]: DEBUG nova.virt.hardware [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.398151] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b859a1-684c-44ce-a4c9-a124fed31e05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.411217] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f74a014-021b-4ab8-9945-f3e1f05fd81e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.434044] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.117s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.434044] env[61986]: ERROR nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Traceback (most recent call last): [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self.driver.spawn(context, instance, image_meta, [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.434044] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] vm_ref = self.build_virtual_machine(instance, [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] for vif in network_info: [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return self._sync_wrapper(fn, *args, **kwargs) [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self.wait() [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self[:] = self._gt.wait() [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return self._exit_event.wait() [ 651.434633] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] result = hub.switch() [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return self.greenlet.switch() [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] result = function(*args, **kwargs) [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] return func(*args, **kwargs) [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] raise e [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] nwinfo = self.network_api.allocate_for_instance( [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.436030] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] created_port_ids = self._update_ports_for_instance( [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] with excutils.save_and_reraise_exception(): [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] self.force_reraise() [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] raise self.value [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] updated_port = self._update_port( [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] _ensure_no_port_binding_failure(port) [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.436734] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] raise exception.PortBindingFailed(port_id=port['id']) [ 651.437227] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] nova.exception.PortBindingFailed: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. [ 651.437227] env[61986]: ERROR nova.compute.manager [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] [ 651.437227] env[61986]: DEBUG nova.compute.utils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 651.437227] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.584s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.441870] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Build of instance ab392d78-a11d-44d0-81d6-8afcb169e6e1 was re-scheduled: Binding failed for port 8a83e57b-4d3f-47cb-a0e4-93fef82d732c, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 651.442819] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 651.442819] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquiring lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.442819] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Acquired lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.447103] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 651.768439] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.965598] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.164371] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.250386] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquiring lock "7e5fadf3-dd41-402c-9878-2d11295203ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.250687] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Lock "7e5fadf3-dd41-402c-9878-2d11295203ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.422203] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df26cfd4-5721-467e-8359-b3c9de6a0833 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.431361] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6392a67-0251-43bb-9ff0-0ce3c417450c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.466538] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cee3e8-fefa-4ad0-9abd-20e4ae72f236 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.476018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcd6756-d5bb-4e41-8ec3-d4de78af5bec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.489476] env[61986]: DEBUG nova.compute.provider_tree [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.667193] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Releasing lock "refresh_cache-ab392d78-a11d-44d0-81d6-8afcb169e6e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.667193] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 652.667678] env[61986]: DEBUG nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.667678] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 652.698750] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.869113] env[61986]: DEBUG nova.compute.manager [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Received event network-changed-d5b88187-18fc-4690-9ab8-e8c53d1a6a95 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 652.869497] env[61986]: DEBUG nova.compute.manager [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Refreshing instance network info cache due to event network-changed-d5b88187-18fc-4690-9ab8-e8c53d1a6a95. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 652.871722] env[61986]: DEBUG oslo_concurrency.lockutils [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] Acquiring lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.871722] env[61986]: DEBUG oslo_concurrency.lockutils [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] Acquired lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.871722] env[61986]: DEBUG nova.network.neutron [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Refreshing network info cache for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 652.950108] env[61986]: ERROR nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 652.950108] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.950108] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.950108] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.950108] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.950108] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.950108] env[61986]: ERROR nova.compute.manager raise self.value [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.950108] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.950108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.950108] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.950797] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.950797] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.950797] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 652.950797] env[61986]: ERROR nova.compute.manager [ 652.950797] env[61986]: Traceback (most recent call last): [ 652.950797] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.950797] env[61986]: listener.cb(fileno) [ 652.950797] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.950797] env[61986]: result = function(*args, **kwargs) [ 652.950797] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.950797] env[61986]: return func(*args, **kwargs) [ 652.950797] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.950797] env[61986]: raise e [ 652.950797] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.950797] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 652.950797] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.950797] env[61986]: created_port_ids = self._update_ports_for_instance( [ 652.950797] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.950797] env[61986]: with excutils.save_and_reraise_exception(): [ 652.950797] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.950797] env[61986]: self.force_reraise() [ 652.950797] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.950797] env[61986]: raise self.value [ 652.950797] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.950797] env[61986]: updated_port = self._update_port( [ 652.950797] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.950797] env[61986]: _ensure_no_port_binding_failure(port) [ 652.950797] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.950797] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.951918] env[61986]: nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 652.951918] env[61986]: Removing descriptor: 14 [ 652.951918] env[61986]: ERROR nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Traceback (most recent call last): [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] yield resources [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self.driver.spawn(context, instance, image_meta, [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.951918] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] vm_ref = self.build_virtual_machine(instance, [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] for vif in network_info: [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return self._sync_wrapper(fn, *args, **kwargs) [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self.wait() [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self[:] = self._gt.wait() [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return self._exit_event.wait() [ 652.952363] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] result = hub.switch() [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return self.greenlet.switch() [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] result = function(*args, **kwargs) [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return func(*args, **kwargs) [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] raise e [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] nwinfo = self.network_api.allocate_for_instance( [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.952822] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] created_port_ids = self._update_ports_for_instance( [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] with excutils.save_and_reraise_exception(): [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self.force_reraise() [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] raise self.value [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] updated_port = self._update_port( [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] _ensure_no_port_binding_failure(port) [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.953534] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] raise exception.PortBindingFailed(port_id=port['id']) [ 652.954704] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 652.954704] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] [ 652.954704] env[61986]: INFO nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Terminating instance [ 652.959030] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquiring lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.995816] env[61986]: DEBUG nova.scheduler.client.report [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.203148] env[61986]: DEBUG nova.network.neutron [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.396160] env[61986]: DEBUG nova.network.neutron [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.500186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.065s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.500830] env[61986]: ERROR nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Traceback (most recent call last): [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self.driver.spawn(context, instance, image_meta, [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] vm_ref = self.build_virtual_machine(instance, [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.500830] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] for vif in network_info: [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return self._sync_wrapper(fn, *args, **kwargs) [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self.wait() [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self[:] = self._gt.wait() [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return self._exit_event.wait() [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] result = hub.switch() [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.501309] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return self.greenlet.switch() [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] result = function(*args, **kwargs) [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] return func(*args, **kwargs) [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] raise e [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] nwinfo = self.network_api.allocate_for_instance( [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] created_port_ids = self._update_ports_for_instance( [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] with excutils.save_and_reraise_exception(): [ 653.501782] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] self.force_reraise() [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] raise self.value [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] updated_port = self._update_port( [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] _ensure_no_port_binding_failure(port) [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] raise exception.PortBindingFailed(port_id=port['id']) [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] nova.exception.PortBindingFailed: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. [ 653.502246] env[61986]: ERROR nova.compute.manager [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] [ 653.502783] env[61986]: DEBUG nova.compute.utils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.505216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.308s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.505471] env[61986]: DEBUG nova.objects.instance [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lazy-loading 'resources' on Instance uuid 6b586395-eb84-44f6-84c8-4a3ef3d64254 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 653.506978] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Build of instance 40b8df0a-36f1-4a4d-a975-f596901d1f98 was re-scheduled: Binding failed for port e4549b4f-a8bd-4f44-87d4-b5cf94934705, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.507518] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.507637] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquiring lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.507773] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Acquired lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.507922] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 653.577866] env[61986]: DEBUG nova.network.neutron [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.708933] env[61986]: INFO nova.compute.manager [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] [instance: ab392d78-a11d-44d0-81d6-8afcb169e6e1] Took 1.04 seconds to deallocate network for instance. [ 654.038690] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.086611] env[61986]: DEBUG oslo_concurrency.lockutils [req-0462d9e9-b9a6-4ba3-8856-5186b62a6e35 req-0902850b-28f8-4e12-9b2b-029db0fc170f service nova] Releasing lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.088115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquired lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.089933] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.212683] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.450932] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce4d99a-48fa-4ee7-8b62-0fc9faab2e98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.458573] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb30607-7949-4e09-93d1-37d7b7c68efb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.498835] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db11557-fb71-407e-ab73-d861160ecd78 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.513019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088f16c1-d17f-40b0-a37f-8a6a452f4a4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.528502] env[61986]: DEBUG nova.compute.provider_tree [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.624169] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.721329] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Releasing lock "refresh_cache-40b8df0a-36f1-4a4d-a975-f596901d1f98" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.721550] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.722162] env[61986]: DEBUG nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.722240] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 654.749974] env[61986]: INFO nova.scheduler.client.report [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Deleted allocations for instance ab392d78-a11d-44d0-81d6-8afcb169e6e1 [ 654.757548] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.810799] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.035219] env[61986]: DEBUG nova.scheduler.client.report [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.114445] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "198f36d0-8be4-4885-9c0b-e85d6a0cbb22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.114693] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "198f36d0-8be4-4885-9c0b-e85d6a0cbb22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.264710] env[61986]: DEBUG nova.network.neutron [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.266285] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f3d030d1-ca67-4f7b-8426-1da35f72a44b tempest-VolumesAssistedSnapshotsTest-1572619058 tempest-VolumesAssistedSnapshotsTest-1572619058-project-member] Lock "ab392d78-a11d-44d0-81d6-8afcb169e6e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.606s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.313384] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Releasing lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.313922] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.314131] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 655.314433] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d943d991-ae46-493e-930a-47194a2bb4d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.329951] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314af1c6-4cb8-4ea5-8906-25c207a4feb6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.360108] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49d08e2b-1be4-4adb-98a8-98efad8eb7ff could not be found. [ 655.360717] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.360717] env[61986]: INFO nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Took 0.05 seconds to destroy the instance on the hypervisor. [ 655.361088] env[61986]: DEBUG oslo.service.loopingcall [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.361189] env[61986]: DEBUG nova.compute.manager [-] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.363317] env[61986]: DEBUG nova.network.neutron [-] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.389791] env[61986]: DEBUG nova.network.neutron [-] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.509912] env[61986]: DEBUG nova.compute.manager [req-0fd7f82b-d1f1-4a6f-bacc-b9099f3fd7cd req-61a670cf-32ea-4cd2-9cda-8cf198946a7b service nova] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Received event network-vif-deleted-d5b88187-18fc-4690-9ab8-e8c53d1a6a95 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.543568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.551018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.377s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.551018] env[61986]: DEBUG nova.objects.instance [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 655.565467] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "77f4d9bd-c758-44c3-a17c-1efa8c395eb1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.566724] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "77f4d9bd-c758-44c3-a17c-1efa8c395eb1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.583379] env[61986]: INFO nova.scheduler.client.report [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Deleted allocations for instance 6b586395-eb84-44f6-84c8-4a3ef3d64254 [ 655.769322] env[61986]: INFO nova.compute.manager [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] [instance: 40b8df0a-36f1-4a4d-a975-f596901d1f98] Took 1.05 seconds to deallocate network for instance. [ 655.771823] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 655.895648] env[61986]: DEBUG nova.network.neutron [-] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.093377] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d9cb798-e2cb-4adf-950f-3472b43ac974 tempest-ServerDiagnosticsV248Test-1926246162 tempest-ServerDiagnosticsV248Test-1926246162-project-member] Lock "6b586395-eb84-44f6-84c8-4a3ef3d64254" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.949s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.321370] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.398129] env[61986]: INFO nova.compute.manager [-] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Took 1.04 seconds to deallocate network for instance. [ 656.400719] env[61986]: DEBUG nova.compute.claims [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 656.400910] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.557845] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e98d29cd-fdf9-4ff8-a865-86330e90c75e tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.558874] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.269s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.560921] env[61986]: INFO nova.compute.claims [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.822013] env[61986]: INFO nova.scheduler.client.report [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Deleted allocations for instance 40b8df0a-36f1-4a4d-a975-f596901d1f98 [ 657.333072] env[61986]: DEBUG oslo_concurrency.lockutils [None req-58c4063c-a257-49af-8cef-278e915f3374 tempest-ServerRescueNegativeTestJSON-756372731 tempest-ServerRescueNegativeTestJSON-756372731-project-member] Lock "40b8df0a-36f1-4a4d-a975-f596901d1f98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.394s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.847110] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.007562] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049557f0-7223-403a-be2f-c80d1007b894 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.021333] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128a9b14-abc1-41a8-9bbb-1dd1a3e60d53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.052708] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462252bf-0359-4efb-b0d3-fcd13926f385 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.060221] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5fa930-bfad-48a0-98e6-66fa0d4b0889 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.075393] env[61986]: DEBUG nova.compute.provider_tree [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.382635] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.582670] env[61986]: DEBUG nova.scheduler.client.report [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.090040] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.090040] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.096379] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.299s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.598687] env[61986]: DEBUG nova.compute.utils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.600131] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.600304] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.753599] env[61986]: DEBUG nova.policy [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6881bdd48c4f0ea6bcb7d2d3746d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e9d0c35a4de4f5f9829a7f3c88fde92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 660.014335] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9af2d81-63ba-4dbf-a8b0-18de34fb560f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.024599] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f37bc8-3b7a-4eae-8973-36e4fdc2bd62 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.064593] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3e45fd-204d-4e28-9b97-d6c70cdbd329 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.077028] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8570176-b750-43f9-8e6d-eac13633b747 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.091302] env[61986]: DEBUG nova.compute.provider_tree [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.103800] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.597504] env[61986]: DEBUG nova.scheduler.client.report [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.103292] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.103716] env[61986]: ERROR nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Traceback (most recent call last): [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self.driver.spawn(context, instance, image_meta, [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] vm_ref = self.build_virtual_machine(instance, [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.103716] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] for vif in network_info: [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return self._sync_wrapper(fn, *args, **kwargs) [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self.wait() [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self[:] = self._gt.wait() [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return self._exit_event.wait() [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] result = hub.switch() [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.104124] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return self.greenlet.switch() [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] result = function(*args, **kwargs) [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] return func(*args, **kwargs) [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] raise e [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] nwinfo = self.network_api.allocate_for_instance( [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] created_port_ids = self._update_ports_for_instance( [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] with excutils.save_and_reraise_exception(): [ 661.104534] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] self.force_reraise() [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] raise self.value [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] updated_port = self._update_port( [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] _ensure_no_port_binding_failure(port) [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] raise exception.PortBindingFailed(port_id=port['id']) [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] nova.exception.PortBindingFailed: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. [ 661.104928] env[61986]: ERROR nova.compute.manager [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] [ 661.105395] env[61986]: DEBUG nova.compute.utils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.106327] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.136s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.109996] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Build of instance e51ac29e-48f7-4589-ab67-c83a616cc591 was re-scheduled: Binding failed for port 6dcde07b-d315-4bbd-bdc6-6ae9746f4dc0, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.112285] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.112667] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquiring lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.112893] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Acquired lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.113884] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.121566] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.148653] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.148914] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.149083] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.149267] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.149407] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.149547] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.149746] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.149900] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.150525] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.150841] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.151145] env[61986]: DEBUG nova.virt.hardware [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.152696] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b2750b-3d8e-45eb-acb0-a607436cbb0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.158423] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Successfully created port: ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.171264] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a8c40e-fca1-4668-ae74-6b0b608d11f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.649272] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.850455] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.124041] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20624a1-4dc4-4b0b-9ea7-2d2fff0d1d53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.134828] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4d526c-bab5-4afe-ac6a-fdac69122dc1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.180571] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8081ab5-8168-4bcf-8e38-0392193e27e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.189497] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe25c30-7cd0-413c-94c8-54c7a6752fba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.205207] env[61986]: DEBUG nova.compute.provider_tree [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.354012] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Releasing lock "refresh_cache-e51ac29e-48f7-4589-ab67-c83a616cc591" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.354119] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.354584] env[61986]: DEBUG nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.354584] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.392925] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.538346] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquiring lock "55102fef-de7c-4505-a192-39568e08c096" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.540582] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Lock "55102fef-de7c-4505-a192-39568e08c096" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.711254] env[61986]: DEBUG nova.scheduler.client.report [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.896987] env[61986]: DEBUG nova.network.neutron [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.214758] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.108s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.215735] env[61986]: ERROR nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Traceback (most recent call last): [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self.driver.spawn(context, instance, image_meta, [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] vm_ref = self.build_virtual_machine(instance, [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.215735] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] for vif in network_info: [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return self._sync_wrapper(fn, *args, **kwargs) [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self.wait() [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self[:] = self._gt.wait() [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return self._exit_event.wait() [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] result = hub.switch() [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.216765] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return self.greenlet.switch() [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] result = function(*args, **kwargs) [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] return func(*args, **kwargs) [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] raise e [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] nwinfo = self.network_api.allocate_for_instance( [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] created_port_ids = self._update_ports_for_instance( [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] with excutils.save_and_reraise_exception(): [ 663.220053] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] self.force_reraise() [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] raise self.value [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] updated_port = self._update_port( [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] _ensure_no_port_binding_failure(port) [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] raise exception.PortBindingFailed(port_id=port['id']) [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] nova.exception.PortBindingFailed: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. [ 663.220503] env[61986]: ERROR nova.compute.manager [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] [ 663.220870] env[61986]: DEBUG nova.compute.utils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.220870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.845s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.220870] env[61986]: DEBUG nova.objects.instance [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 663.223154] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Build of instance 5aa473df-1720-43f3-b3ca-5ab9ff565378 was re-scheduled: Binding failed for port 35badee2-138a-4abb-9304-06fde6862a96, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.225597] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.225774] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.225918] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.226122] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.399532] env[61986]: INFO nova.compute.manager [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] [instance: e51ac29e-48f7-4589-ab67-c83a616cc591] Took 1.04 seconds to deallocate network for instance. [ 663.786558] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.849944] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.055339] env[61986]: ERROR nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 664.055339] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.055339] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.055339] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.055339] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.055339] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.055339] env[61986]: ERROR nova.compute.manager raise self.value [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.055339] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.055339] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.055339] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.056068] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.056068] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.056068] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 664.056068] env[61986]: ERROR nova.compute.manager [ 664.056068] env[61986]: Traceback (most recent call last): [ 664.056068] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.056068] env[61986]: listener.cb(fileno) [ 664.056068] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.056068] env[61986]: result = function(*args, **kwargs) [ 664.056068] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.056068] env[61986]: return func(*args, **kwargs) [ 664.056068] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.056068] env[61986]: raise e [ 664.056068] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.056068] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 664.056068] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.056068] env[61986]: created_port_ids = self._update_ports_for_instance( [ 664.056068] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.056068] env[61986]: with excutils.save_and_reraise_exception(): [ 664.056068] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.056068] env[61986]: self.force_reraise() [ 664.056068] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.056068] env[61986]: raise self.value [ 664.056068] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.056068] env[61986]: updated_port = self._update_port( [ 664.056068] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.056068] env[61986]: _ensure_no_port_binding_failure(port) [ 664.056068] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.056068] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.057105] env[61986]: nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 664.057105] env[61986]: Removing descriptor: 14 [ 664.057105] env[61986]: ERROR nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Traceback (most recent call last): [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] yield resources [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self.driver.spawn(context, instance, image_meta, [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.057105] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] vm_ref = self.build_virtual_machine(instance, [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] for vif in network_info: [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return self._sync_wrapper(fn, *args, **kwargs) [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self.wait() [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self[:] = self._gt.wait() [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return self._exit_event.wait() [ 664.057531] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] result = hub.switch() [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return self.greenlet.switch() [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] result = function(*args, **kwargs) [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return func(*args, **kwargs) [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] raise e [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] nwinfo = self.network_api.allocate_for_instance( [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.057961] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] created_port_ids = self._update_ports_for_instance( [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] with excutils.save_and_reraise_exception(): [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self.force_reraise() [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] raise self.value [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] updated_port = self._update_port( [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] _ensure_no_port_binding_failure(port) [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.058384] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] raise exception.PortBindingFailed(port_id=port['id']) [ 664.058857] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 664.058857] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] [ 664.058857] env[61986]: INFO nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Terminating instance [ 664.059598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.060685] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.060685] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 664.238981] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fbe39af-6608-4426-83a7-413d060f8552 tempest-ServersAdmin275Test-864818061 tempest-ServersAdmin275Test-864818061-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.242916] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.669s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.355068] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-5aa473df-1720-43f3-b3ca-5ab9ff565378" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.355068] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.355068] env[61986]: DEBUG nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.355068] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.384233] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.441145] env[61986]: INFO nova.scheduler.client.report [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Deleted allocations for instance e51ac29e-48f7-4589-ab67-c83a616cc591 [ 664.458577] env[61986]: DEBUG nova.compute.manager [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Received event network-changed-ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.458577] env[61986]: DEBUG nova.compute.manager [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Refreshing instance network info cache due to event network-changed-ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 664.458577] env[61986]: DEBUG oslo_concurrency.lockutils [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] Acquiring lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.605089] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.791625] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.887231] env[61986]: DEBUG nova.network.neutron [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.953991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-378a69c8-b252-4395-8f27-72611dda296e tempest-ServersWithSpecificFlavorTestJSON-809754255 tempest-ServersWithSpecificFlavorTestJSON-809754255-project-member] Lock "e51ac29e-48f7-4589-ab67-c83a616cc591" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.184s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.060371] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "8bc12f2b-a632-4ffe-9d55-1a838e94931b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.060371] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "8bc12f2b-a632-4ffe-9d55-1a838e94931b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.116621] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "2377b482-52e7-4054-8d5b-8f00587c3371" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.116621] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "2377b482-52e7-4054-8d5b-8f00587c3371" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.295560] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.295560] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.295560] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 665.295947] env[61986]: DEBUG oslo_concurrency.lockutils [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] Acquired lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.296262] env[61986]: DEBUG nova.network.neutron [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Refreshing network info cache for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.298889] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64bc4f26-bdfc-4d80-b52c-5cbafcb6d66f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.315094] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de7b620-1bcf-4198-93cc-b46095b58aa2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.330098] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cca176-b45e-4dfa-b1b8-dc6389e69d6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.340133] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a074e643-7cdd-4d8d-bbff-ee1ec4b59663 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.351324] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 10568cfc-3634-47d7-96d6-893c4fe01282 could not be found. [ 665.351324] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 665.351324] env[61986]: INFO nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Took 0.06 seconds to destroy the instance on the hypervisor. [ 665.351543] env[61986]: DEBUG oslo.service.loopingcall [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.352258] env[61986]: DEBUG nova.compute.manager [-] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.352328] env[61986]: DEBUG nova.network.neutron [-] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 665.388350] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935517a1-c906-45a6-a392-fcedf6653d2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.391450] env[61986]: INFO nova.compute.manager [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 5aa473df-1720-43f3-b3ca-5ab9ff565378] Took 1.04 seconds to deallocate network for instance. [ 665.398960] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70929cf-db04-4f8a-bdb0-1614f622f460 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.414839] env[61986]: DEBUG nova.compute.provider_tree [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.427030] env[61986]: DEBUG nova.network.neutron [-] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.460429] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.507538] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "aa12919e-3891-4da9-a280-4155137864e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.507538] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa12919e-3891-4da9-a280-4155137864e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.838523] env[61986]: DEBUG nova.network.neutron [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.918262] env[61986]: DEBUG nova.scheduler.client.report [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.929300] env[61986]: DEBUG nova.network.neutron [-] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.985928] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.177621] env[61986]: DEBUG nova.network.neutron [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.426247] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.182s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.426247] env[61986]: ERROR nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Traceback (most recent call last): [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self.driver.spawn(context, instance, image_meta, [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.426247] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] vm_ref = self.build_virtual_machine(instance, [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] for vif in network_info: [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return self._sync_wrapper(fn, *args, **kwargs) [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self.wait() [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self[:] = self._gt.wait() [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return self._exit_event.wait() [ 666.426881] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] result = hub.switch() [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return self.greenlet.switch() [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] result = function(*args, **kwargs) [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] return func(*args, **kwargs) [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] raise e [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] nwinfo = self.network_api.allocate_for_instance( [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.427298] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] created_port_ids = self._update_ports_for_instance( [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] with excutils.save_and_reraise_exception(): [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] self.force_reraise() [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] raise self.value [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] updated_port = self._update_port( [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] _ensure_no_port_binding_failure(port) [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.427935] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] raise exception.PortBindingFailed(port_id=port['id']) [ 666.428340] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] nova.exception.PortBindingFailed: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. [ 666.428340] env[61986]: ERROR nova.compute.manager [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] [ 666.428340] env[61986]: DEBUG nova.compute.utils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.428340] env[61986]: INFO nova.scheduler.client.report [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Deleted allocations for instance 5aa473df-1720-43f3-b3ca-5ab9ff565378 [ 666.434869] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.089s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.436422] env[61986]: INFO nova.compute.manager [-] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Took 1.08 seconds to deallocate network for instance. [ 666.436947] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Build of instance f5af7037-96fd-49f1-af80-a528737a3fa7 was re-scheduled: Binding failed for port 9686a128-bc3c-4094-9517-e0cfcd644c32, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.437410] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.437624] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquiring lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.437759] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Acquired lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.437909] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.443306] env[61986]: DEBUG nova.compute.claims [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 666.443306] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.680672] env[61986]: DEBUG oslo_concurrency.lockutils [req-3bc23f6c-2c04-401f-b369-b52a054764a7 req-d2de013a-f0f2-4d02-8873-ae42a8dee896 service nova] Releasing lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.907480] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.907548] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.947688] env[61986]: DEBUG oslo_concurrency.lockutils [None req-15be9848-f11f-4f98-988d-e33a15847be6 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "5aa473df-1720-43f3-b3ca-5ab9ff565378" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.978s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.966633] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.068102] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.307169] env[61986]: DEBUG nova.compute.manager [req-dec63959-acd8-436e-9b4c-30307a846101 req-c54a8da5-3c98-4d51-a905-2ea4e960686c service nova] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Received event network-vif-deleted-ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.439098] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfb8734-0633-4ac3-9804-2d1d82fa5605 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.448836] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338b875e-1202-4e17-8c31-db2b0d161fed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.452630] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.483571] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a58bec7-1768-4823-903f-ca9e4237885d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.492028] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc38be7-0b1a-441c-8bf0-9aaf443c648b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.507488] env[61986]: DEBUG nova.compute.provider_tree [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.571260] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Releasing lock "refresh_cache-f5af7037-96fd-49f1-af80-a528737a3fa7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.571514] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.571693] env[61986]: DEBUG nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.571843] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.595570] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.987254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.010688] env[61986]: DEBUG nova.scheduler.client.report [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.102651] env[61986]: DEBUG nova.network.neutron [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.516947] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.083s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.518883] env[61986]: ERROR nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] Traceback (most recent call last): [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self.driver.spawn(context, instance, image_meta, [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] vm_ref = self.build_virtual_machine(instance, [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.518883] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] for vif in network_info: [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return self._sync_wrapper(fn, *args, **kwargs) [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self.wait() [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self[:] = self._gt.wait() [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return self._exit_event.wait() [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] result = hub.switch() [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.519325] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return self.greenlet.switch() [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] result = function(*args, **kwargs) [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] return func(*args, **kwargs) [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] raise e [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] nwinfo = self.network_api.allocate_for_instance( [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] created_port_ids = self._update_ports_for_instance( [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] with excutils.save_and_reraise_exception(): [ 668.519756] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] self.force_reraise() [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] raise self.value [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] updated_port = self._update_port( [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] _ensure_no_port_binding_failure(port) [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] raise exception.PortBindingFailed(port_id=port['id']) [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] nova.exception.PortBindingFailed: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. [ 668.520460] env[61986]: ERROR nova.compute.manager [instance: 42698792-0555-45ae-a525-62265259bc17] [ 668.520815] env[61986]: DEBUG nova.compute.utils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.523193] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Build of instance 42698792-0555-45ae-a525-62265259bc17 was re-scheduled: Binding failed for port 42873233-21b5-4ad3-a938-075026f9859e, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.523707] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.523880] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.524041] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.524249] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.525374] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.590s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.578114] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "6c62a1e9-d60c-4720-a554-951bea4b2e18" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.578542] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.605695] env[61986]: INFO nova.compute.manager [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] [instance: f5af7037-96fd-49f1-af80-a528737a3fa7] Took 1.03 seconds to deallocate network for instance. [ 669.068171] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.225848] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.526554] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5578ca47-5a2c-4722-a8de-d9e8d3c59753 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.535813] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ec2633-6e4e-4c78-9330-c26da24ce318 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.567126] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cd86a8-e97d-4f08-920d-6e74012fee43 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.574708] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47a4c73-0d74-4730-805e-b46640c77ce9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.589426] env[61986]: DEBUG nova.compute.provider_tree [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.647712] env[61986]: INFO nova.scheduler.client.report [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Deleted allocations for instance f5af7037-96fd-49f1-af80-a528737a3fa7 [ 669.729938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-42698792-0555-45ae-a525-62265259bc17" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.730388] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.730580] env[61986]: DEBUG nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.730744] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.733210] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "40de5f22-0de2-466a-91ab-dcb6ec586dad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.733435] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.747851] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.092072] env[61986]: DEBUG nova.scheduler.client.report [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.160286] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43853b06-ae44-4a01-813f-6adf021ad98e tempest-InstanceActionsNegativeTestJSON-1943113433 tempest-InstanceActionsNegativeTestJSON-1943113433-project-member] Lock "f5af7037-96fd-49f1-af80-a528737a3fa7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.020s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.251358] env[61986]: DEBUG nova.network.neutron [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.597333] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.598052] env[61986]: ERROR nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Traceback (most recent call last): [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self.driver.spawn(context, instance, image_meta, [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] vm_ref = self.build_virtual_machine(instance, [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.598052] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] for vif in network_info: [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return self._sync_wrapper(fn, *args, **kwargs) [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self.wait() [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self[:] = self._gt.wait() [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return self._exit_event.wait() [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] result = hub.switch() [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.598532] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return self.greenlet.switch() [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] result = function(*args, **kwargs) [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] return func(*args, **kwargs) [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] raise e [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] nwinfo = self.network_api.allocate_for_instance( [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] created_port_ids = self._update_ports_for_instance( [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] with excutils.save_and_reraise_exception(): [ 670.598941] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] self.force_reraise() [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] raise self.value [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] updated_port = self._update_port( [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] _ensure_no_port_binding_failure(port) [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] raise exception.PortBindingFailed(port_id=port['id']) [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] nova.exception.PortBindingFailed: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. [ 670.599368] env[61986]: ERROR nova.compute.manager [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] [ 670.599728] env[61986]: DEBUG nova.compute.utils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.600599] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.330s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.600821] env[61986]: DEBUG nova.objects.instance [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lazy-loading 'resources' on Instance uuid 76b7db85-734f-40fe-8d58-2fba89bf205f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 670.603477] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Build of instance c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a was re-scheduled: Binding failed for port b5d913fb-250b-4705-b9f6-35103c57c660, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.603477] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.603667] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquiring lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.603710] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Acquired lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.603865] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.665180] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.757878] env[61986]: INFO nova.compute.manager [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 42698792-0555-45ae-a525-62265259bc17] Took 1.02 seconds to deallocate network for instance. [ 671.167033] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.195872] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.537056] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.639576] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7090e248-e5aa-460e-81e2-f011cf8bde60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.649942] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46b2f46-a207-4419-9dee-6ffba94ffcf9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.683363] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bc9029-4b5d-49d5-9a24-318e247e1085 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.690228] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c0a578-0ff5-4ddb-a54e-5bc9307eea19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.702992] env[61986]: DEBUG nova.compute.provider_tree [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.798388] env[61986]: INFO nova.scheduler.client.report [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted allocations for instance 42698792-0555-45ae-a525-62265259bc17 [ 672.041598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Releasing lock "refresh_cache-c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.041833] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.042415] env[61986]: DEBUG nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.042415] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.081038] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.214225] env[61986]: DEBUG nova.scheduler.client.report [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.311933] env[61986]: DEBUG oslo_concurrency.lockutils [None req-45230147-7a26-4721-bca1-97d95004201b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "42698792-0555-45ae-a525-62265259bc17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.214s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.582656] env[61986]: DEBUG nova.network.neutron [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.718917] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.118s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.725074] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.954s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.726693] env[61986]: INFO nova.compute.claims [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.757444] env[61986]: INFO nova.scheduler.client.report [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Deleted allocations for instance 76b7db85-734f-40fe-8d58-2fba89bf205f [ 672.817572] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.089482] env[61986]: INFO nova.compute.manager [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] [instance: c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a] Took 1.05 seconds to deallocate network for instance. [ 673.270074] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9693c7f2-7cec-40eb-a92b-337dfe880ad2 tempest-ServersAdmin275Test-2001703186 tempest-ServersAdmin275Test-2001703186-project-member] Lock "76b7db85-734f-40fe-8d58-2fba89bf205f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.821s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.357185] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.959596] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.959596] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.125027] env[61986]: INFO nova.scheduler.client.report [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Deleted allocations for instance c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a [ 674.219341] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b4c4f5-f6e0-43a9-9c87-7838a98ccd26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.236629] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00338fd9-6136-4ce1-8d8b-9215b8e2ecd0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.275304] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01622e2-9b90-4858-a88d-3025f7e8ac06 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.284311] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b25eec6-26cb-401b-b227-b9c4399eb229 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.299698] env[61986]: DEBUG nova.compute.provider_tree [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.631504] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84cb2765-41e8-474c-aef7-2640d8187f4b tempest-DeleteServersAdminTestJSON-199322564 tempest-DeleteServersAdminTestJSON-199322564-project-member] Lock "c3960b9d-e8b9-4a3e-8fb2-2a6218d4886a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.422s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.803190] env[61986]: DEBUG nova.scheduler.client.report [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.135608] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.309907] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.310494] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.313474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.994s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.314958] env[61986]: INFO nova.compute.claims [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.666038] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.821640] env[61986]: DEBUG nova.compute.utils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.825177] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.825398] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.927892] env[61986]: DEBUG nova.policy [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ea19b4c355d4b2797683d78e82eadea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0110dfcd30de444badf6fe685042ba3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 676.308350] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.308780] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "977799b4-2793-4513-9447-483146fc7ac4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.325818] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.503778] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Successfully created port: 255384ec-de5e-4e9e-8371-c22f2e0f32fa {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.840582] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f76ebd-76de-4067-a110-4fc6c6e107ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.850022] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0d3c8b-385b-4c96-87c8-6b01733f28af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.887277] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e1ebf2-7b85-42ff-a6cd-d0921386b463 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.895849] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308bf3ca-f87e-4236-8a7b-b729a8c98051 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.911284] env[61986]: DEBUG nova.compute.provider_tree [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.029792] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "4bb588b3-c5e1-4619-b77b-6734d6516170" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.029792] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.342017] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.420874] env[61986]: DEBUG nova.scheduler.client.report [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.467433] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.467433] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.467433] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.468023] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.468533] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.468801] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.469118] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.469374] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.469647] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.469899] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.470199] env[61986]: DEBUG nova.virt.hardware [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.471173] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9fc051-cece-491c-bdf5-58ed0658794e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.499274] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773c0e83-3567-41b5-9298-6f434adbe619 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.926470] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.927296] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.931163] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.530s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.441468] env[61986]: DEBUG nova.compute.utils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.450694] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.450901] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 678.532696] env[61986]: DEBUG nova.policy [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34bc65b11c324fa9b24b58488ec6a902', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5b11c088c3c4448283e1f35c1370a8e6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 678.628581] env[61986]: ERROR nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 678.628581] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.628581] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 678.628581] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 678.628581] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.628581] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.628581] env[61986]: ERROR nova.compute.manager raise self.value [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 678.628581] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 678.628581] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.628581] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 678.629047] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.629047] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 678.629047] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 678.629047] env[61986]: ERROR nova.compute.manager [ 678.629047] env[61986]: Traceback (most recent call last): [ 678.629047] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 678.629047] env[61986]: listener.cb(fileno) [ 678.629047] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.629047] env[61986]: result = function(*args, **kwargs) [ 678.629047] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.629047] env[61986]: return func(*args, **kwargs) [ 678.629047] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.629047] env[61986]: raise e [ 678.629047] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.629047] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 678.629047] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 678.629047] env[61986]: created_port_ids = self._update_ports_for_instance( [ 678.629047] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 678.629047] env[61986]: with excutils.save_and_reraise_exception(): [ 678.629047] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.629047] env[61986]: self.force_reraise() [ 678.629047] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.629047] env[61986]: raise self.value [ 678.629047] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 678.629047] env[61986]: updated_port = self._update_port( [ 678.629047] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.629047] env[61986]: _ensure_no_port_binding_failure(port) [ 678.629047] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.629047] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 678.629738] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 678.629738] env[61986]: Removing descriptor: 14 [ 678.629738] env[61986]: ERROR nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Traceback (most recent call last): [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] yield resources [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self.driver.spawn(context, instance, image_meta, [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.629738] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] vm_ref = self.build_virtual_machine(instance, [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] for vif in network_info: [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return self._sync_wrapper(fn, *args, **kwargs) [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self.wait() [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self[:] = self._gt.wait() [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return self._exit_event.wait() [ 678.630049] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] result = hub.switch() [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return self.greenlet.switch() [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] result = function(*args, **kwargs) [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return func(*args, **kwargs) [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] raise e [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] nwinfo = self.network_api.allocate_for_instance( [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 678.630366] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] created_port_ids = self._update_ports_for_instance( [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] with excutils.save_and_reraise_exception(): [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self.force_reraise() [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] raise self.value [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] updated_port = self._update_port( [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] _ensure_no_port_binding_failure(port) [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.630661] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] raise exception.PortBindingFailed(port_id=port['id']) [ 678.631028] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 678.631028] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] [ 678.631028] env[61986]: INFO nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Terminating instance [ 678.632153] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.632244] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.632396] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.949420] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.956552] env[61986]: DEBUG nova.compute.manager [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Received event network-changed-255384ec-de5e-4e9e-8371-c22f2e0f32fa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 678.959601] env[61986]: DEBUG nova.compute.manager [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Refreshing instance network info cache due to event network-changed-255384ec-de5e-4e9e-8371-c22f2e0f32fa. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 678.959601] env[61986]: DEBUG oslo_concurrency.lockutils [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] Acquiring lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.979920] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.979920] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.081784] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5c787e-7e3c-4332-b7c8-70171921bb29 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.090316] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351018e4-a44e-4ead-a4b0-5f3d8a2ba278 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.123673] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Successfully created port: 0a254c6b-1159-4515-ad39-1bd51a9624aa {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.126248] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b022ed7-223a-46b3-8483-da20aeea4b41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.134182] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc37e7ad-61cf-4641-b165-f6e52a14afc8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.149597] env[61986]: DEBUG nova.compute.provider_tree [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.167730] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.293570] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.656677] env[61986]: DEBUG nova.scheduler.client.report [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.796191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.796643] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.796878] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 679.797218] env[61986]: DEBUG oslo_concurrency.lockutils [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] Acquired lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.797389] env[61986]: DEBUG nova.network.neutron [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Refreshing network info cache for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 679.798513] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54434493-c09f-4a7e-a712-c5d9a140346a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.812986] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969c8602-5f4b-45a2-bf33-35edf983aebf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.837085] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 236ecde7-1250-4ce1-837c-8266b95dd513 could not be found. [ 679.837919] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 679.838151] env[61986]: INFO nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Took 0.04 seconds to destroy the instance on the hypervisor. [ 679.838400] env[61986]: DEBUG oslo.service.loopingcall [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.838618] env[61986]: DEBUG nova.compute.manager [-] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.838709] env[61986]: DEBUG nova.network.neutron [-] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.866718] env[61986]: DEBUG nova.network.neutron [-] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.959811] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.991128] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.991453] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.991709] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.991709] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.991813] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.993336] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.993431] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.993964] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.994064] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.994302] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.994905] env[61986]: DEBUG nova.virt.hardware [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.996326] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b2579d-3289-4c57-a971-add33d858c6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.006412] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c05915-48b1-4aea-ab4b-32067a83bd2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.164396] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.232s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.164396] env[61986]: ERROR nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Traceback (most recent call last): [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self.driver.spawn(context, instance, image_meta, [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.164396] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] vm_ref = self.build_virtual_machine(instance, [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] for vif in network_info: [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return self._sync_wrapper(fn, *args, **kwargs) [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self.wait() [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self[:] = self._gt.wait() [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return self._exit_event.wait() [ 680.164764] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] result = hub.switch() [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return self.greenlet.switch() [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] result = function(*args, **kwargs) [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] return func(*args, **kwargs) [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] raise e [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] nwinfo = self.network_api.allocate_for_instance( [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.165103] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] created_port_ids = self._update_ports_for_instance( [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] with excutils.save_and_reraise_exception(): [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] self.force_reraise() [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] raise self.value [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] updated_port = self._update_port( [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] _ensure_no_port_binding_failure(port) [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.165426] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] raise exception.PortBindingFailed(port_id=port['id']) [ 680.165788] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] nova.exception.PortBindingFailed: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. [ 680.165788] env[61986]: ERROR nova.compute.manager [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] [ 680.165788] env[61986]: DEBUG nova.compute.utils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.166875] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Build of instance 49d08e2b-1be4-4adb-98a8-98efad8eb7ff was re-scheduled: Binding failed for port d5b88187-18fc-4690-9ab8-e8c53d1a6a95, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.167885] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.168142] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquiring lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.168256] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Acquired lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.168587] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.171481] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.789s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.174523] env[61986]: INFO nova.compute.claims [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.328189] env[61986]: DEBUG nova.network.neutron [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.372438] env[61986]: DEBUG nova.network.neutron [-] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.470644] env[61986]: DEBUG nova.network.neutron [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.641760] env[61986]: DEBUG nova.compute.manager [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Received event network-changed-0a254c6b-1159-4515-ad39-1bd51a9624aa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 680.642033] env[61986]: DEBUG nova.compute.manager [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Refreshing instance network info cache due to event network-changed-0a254c6b-1159-4515-ad39-1bd51a9624aa. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 680.642648] env[61986]: DEBUG oslo_concurrency.lockutils [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] Acquiring lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.642809] env[61986]: DEBUG oslo_concurrency.lockutils [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] Acquired lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.642975] env[61986]: DEBUG nova.network.neutron [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Refreshing network info cache for port 0a254c6b-1159-4515-ad39-1bd51a9624aa {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.712194] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.830030] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.879034] env[61986]: INFO nova.compute.manager [-] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Took 1.04 seconds to deallocate network for instance. [ 680.880844] env[61986]: DEBUG nova.compute.claims [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 680.881062] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.950019] env[61986]: ERROR nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 680.950019] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.950019] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.950019] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.950019] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.950019] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.950019] env[61986]: ERROR nova.compute.manager raise self.value [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.950019] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.950019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.950019] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.950582] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.950582] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.950582] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 680.950582] env[61986]: ERROR nova.compute.manager [ 680.950582] env[61986]: Traceback (most recent call last): [ 680.950582] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.950582] env[61986]: listener.cb(fileno) [ 680.950582] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.950582] env[61986]: result = function(*args, **kwargs) [ 680.950582] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.950582] env[61986]: return func(*args, **kwargs) [ 680.950582] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.950582] env[61986]: raise e [ 680.950582] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.950582] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 680.950582] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.950582] env[61986]: created_port_ids = self._update_ports_for_instance( [ 680.950582] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.950582] env[61986]: with excutils.save_and_reraise_exception(): [ 680.950582] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.950582] env[61986]: self.force_reraise() [ 680.950582] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.950582] env[61986]: raise self.value [ 680.950582] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.950582] env[61986]: updated_port = self._update_port( [ 680.950582] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.950582] env[61986]: _ensure_no_port_binding_failure(port) [ 680.950582] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.950582] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.951263] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 680.951263] env[61986]: Removing descriptor: 17 [ 680.951263] env[61986]: ERROR nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] Traceback (most recent call last): [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] yield resources [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self.driver.spawn(context, instance, image_meta, [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.951263] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] vm_ref = self.build_virtual_machine(instance, [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] for vif in network_info: [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return self._sync_wrapper(fn, *args, **kwargs) [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self.wait() [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self[:] = self._gt.wait() [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return self._exit_event.wait() [ 680.951522] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] result = hub.switch() [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return self.greenlet.switch() [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] result = function(*args, **kwargs) [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return func(*args, **kwargs) [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] raise e [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] nwinfo = self.network_api.allocate_for_instance( [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.951809] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] created_port_ids = self._update_ports_for_instance( [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] with excutils.save_and_reraise_exception(): [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self.force_reraise() [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] raise self.value [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] updated_port = self._update_port( [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] _ensure_no_port_binding_failure(port) [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.952110] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] raise exception.PortBindingFailed(port_id=port['id']) [ 680.952438] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 680.952438] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] [ 680.952438] env[61986]: INFO nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Terminating instance [ 680.952438] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquiring lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.973749] env[61986]: DEBUG oslo_concurrency.lockutils [req-3abf09b7-81e1-4715-bca8-9ca897967d02 req-a864130f-493a-42cc-a232-4d81547b3739 service nova] Releasing lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.010362] env[61986]: DEBUG nova.compute.manager [req-515a3f92-5d86-4e9e-be1e-c708ed31140a req-55424fd3-6515-41ff-9371-866b54d8addc service nova] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Received event network-vif-deleted-255384ec-de5e-4e9e-8371-c22f2e0f32fa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 681.163583] env[61986]: DEBUG nova.network.neutron [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.249679] env[61986]: DEBUG nova.network.neutron [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.332840] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Releasing lock "refresh_cache-49d08e2b-1be4-4adb-98a8-98efad8eb7ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.332949] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.333158] env[61986]: DEBUG nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.333324] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.349089] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.633015] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b9eb97-ab3c-4ada-9f6e-4cc9078feef8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.646931] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa1bb7c-14e2-45ad-a4d1-7c95e3d4a829 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.687227] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa27c74-ed4b-4dc1-be76-026ce1a05b96 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.698497] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc3e778-7303-4a4e-b88f-f6f3326b86b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.713334] env[61986]: DEBUG nova.compute.provider_tree [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.755143] env[61986]: DEBUG oslo_concurrency.lockutils [req-33d54544-c73b-47ac-8767-b32cdeeb7652 req-b8fb3597-e8c9-4fab-8520-0a0ff8364fb0 service nova] Releasing lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.756644] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquired lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.756644] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.858378] env[61986]: DEBUG nova.network.neutron [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.218202] env[61986]: DEBUG nova.scheduler.client.report [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.266869] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.267130] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.281980] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.361209] env[61986]: INFO nova.compute.manager [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] [instance: 49d08e2b-1be4-4adb-98a8-98efad8eb7ff] Took 1.03 seconds to deallocate network for instance. [ 682.366098] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.668964] env[61986]: DEBUG nova.compute.manager [req-535a1a58-23e5-495b-8fcc-13eb6ee71af9 req-fb6761e8-f673-4dd2-833f-3c02ae69d9cb service nova] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Received event network-vif-deleted-0a254c6b-1159-4515-ad39-1bd51a9624aa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 682.722505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.723598] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.726069] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.740s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.727284] env[61986]: INFO nova.compute.claims [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.868737] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Releasing lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.869123] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 682.869328] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 682.869793] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e8af3c8-f570-43b3-975d-e8659bb5c985 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.879470] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49d4de9-c31e-43b9-826e-4d687e83ff8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.902060] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f70adc74-3b24-434a-829a-13274bae40b1 could not be found. [ 682.902060] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 682.902060] env[61986]: INFO nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 682.902240] env[61986]: DEBUG oslo.service.loopingcall [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.902326] env[61986]: DEBUG nova.compute.manager [-] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.902419] env[61986]: DEBUG nova.network.neutron [-] [instance: f70adc74-3b24-434a-829a-13274bae40b1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.917399] env[61986]: DEBUG nova.network.neutron [-] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.231954] env[61986]: DEBUG nova.compute.utils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.241381] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 683.241381] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 683.287704] env[61986]: DEBUG nova.policy [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '571a9c7054a049bdb698041722fdc204', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ac40bc6ea034e8eb502d7a2b6eb1956', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 683.392267] env[61986]: INFO nova.scheduler.client.report [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Deleted allocations for instance 49d08e2b-1be4-4adb-98a8-98efad8eb7ff [ 683.419810] env[61986]: DEBUG nova.network.neutron [-] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.565929] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Successfully created port: 67d7b46b-c2e9-4979-9ce3-404a43df9063 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.742960] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.900631] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a5fa53f6-ebfd-458f-b8b4-197a8d0f84c5 tempest-ServerAddressesTestJSON-565313581 tempest-ServerAddressesTestJSON-565313581-project-member] Lock "49d08e2b-1be4-4adb-98a8-98efad8eb7ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.112s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.921874] env[61986]: INFO nova.compute.manager [-] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Took 1.02 seconds to deallocate network for instance. [ 683.926511] env[61986]: DEBUG nova.compute.claims [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 683.926511] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.213588] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac58bbed-eed2-4d79-a91d-80d7ac0d6ad0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.213588] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77879b20-7592-42ae-a85d-ed976234a428 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.213588] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83efa390-b89c-47fb-8bbe-05391ff35684 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.213588] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebd2c88-ef2d-49d4-b162-d5e0950c175b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.228963] env[61986]: DEBUG nova.compute.provider_tree [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.404526] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.493309] env[61986]: ERROR nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 684.493309] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.493309] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.493309] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.493309] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.493309] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.493309] env[61986]: ERROR nova.compute.manager raise self.value [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.493309] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.493309] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.493309] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.493774] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.493774] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.493774] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 684.493774] env[61986]: ERROR nova.compute.manager [ 684.493774] env[61986]: Traceback (most recent call last): [ 684.493774] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.493774] env[61986]: listener.cb(fileno) [ 684.493774] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.493774] env[61986]: result = function(*args, **kwargs) [ 684.493774] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.493774] env[61986]: return func(*args, **kwargs) [ 684.493774] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.493774] env[61986]: raise e [ 684.493774] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.493774] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 684.493774] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.493774] env[61986]: created_port_ids = self._update_ports_for_instance( [ 684.493774] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.493774] env[61986]: with excutils.save_and_reraise_exception(): [ 684.493774] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.493774] env[61986]: self.force_reraise() [ 684.493774] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.493774] env[61986]: raise self.value [ 684.493774] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.493774] env[61986]: updated_port = self._update_port( [ 684.493774] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.493774] env[61986]: _ensure_no_port_binding_failure(port) [ 684.493774] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.493774] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.494612] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 684.494612] env[61986]: Removing descriptor: 17 [ 684.696274] env[61986]: DEBUG nova.compute.manager [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Received event network-changed-67d7b46b-c2e9-4979-9ce3-404a43df9063 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 684.696455] env[61986]: DEBUG nova.compute.manager [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Refreshing instance network info cache due to event network-changed-67d7b46b-c2e9-4979-9ce3-404a43df9063. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 684.696659] env[61986]: DEBUG oslo_concurrency.lockutils [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] Acquiring lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.696826] env[61986]: DEBUG oslo_concurrency.lockutils [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] Acquired lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.696943] env[61986]: DEBUG nova.network.neutron [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Refreshing network info cache for port 67d7b46b-c2e9-4979-9ce3-404a43df9063 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 684.731442] env[61986]: DEBUG nova.scheduler.client.report [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.759118] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.785569] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.785841] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.786014] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.786206] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.786350] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.786488] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.788577] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.788808] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.788993] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.789189] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.789360] env[61986]: DEBUG nova.virt.hardware [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.790232] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4609de-ff5f-4120-8b2d-3040e6956195 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.799418] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50288129-bce7-4e32-b2d8-c5f7aff76e95 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.813517] env[61986]: ERROR nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Traceback (most recent call last): [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] yield resources [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self.driver.spawn(context, instance, image_meta, [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] vm_ref = self.build_virtual_machine(instance, [ 684.813517] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] for vif in network_info: [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] return self._sync_wrapper(fn, *args, **kwargs) [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self.wait() [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self[:] = self._gt.wait() [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] return self._exit_event.wait() [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.813884] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] current.throw(*self._exc) [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] result = function(*args, **kwargs) [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] return func(*args, **kwargs) [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] raise e [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] nwinfo = self.network_api.allocate_for_instance( [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] created_port_ids = self._update_ports_for_instance( [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] with excutils.save_and_reraise_exception(): [ 684.814274] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self.force_reraise() [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] raise self.value [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] updated_port = self._update_port( [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] _ensure_no_port_binding_failure(port) [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] raise exception.PortBindingFailed(port_id=port['id']) [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 684.814600] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] [ 684.814600] env[61986]: INFO nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Terminating instance [ 684.816101] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquiring lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.936153] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.232522] env[61986]: DEBUG nova.network.neutron [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.236327] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.236809] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.240141] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.798s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.359754] env[61986]: DEBUG nova.network.neutron [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.741681] env[61986]: DEBUG nova.compute.utils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.743273] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.743480] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.830450] env[61986]: DEBUG nova.policy [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62f168cd4896440e9da69ff7e618a84f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48977d8a09854b8a9dbc1ef24e2ad1c7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 685.862654] env[61986]: DEBUG oslo_concurrency.lockutils [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] Releasing lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.862874] env[61986]: DEBUG nova.compute.manager [req-e67b6006-493d-4cc2-98d1-767657e42f89 req-b6f73ff7-adc1-4ccb-9ce2-4e2f9517010a service nova] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Received event network-vif-deleted-67d7b46b-c2e9-4979-9ce3-404a43df9063 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 685.866830] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquired lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.867114] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 686.186161] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78a0112-af69-4050-8476-536d4fa3faa7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.193942] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5207b56-6cd7-495f-b99b-03f275352873 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.229334] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce044070-980e-4595-add8-4d2bb43a7408 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.237741] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775a4fcd-e956-4136-aa4b-a07c01233638 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.252187] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.255123] env[61986]: DEBUG nova.compute.provider_tree [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.285421] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Successfully created port: 36b058b4-6728-4495-84e4-6059c1ffed46 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.398700] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.520664] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.760971] env[61986]: DEBUG nova.scheduler.client.report [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.996434] env[61986]: DEBUG nova.compute.manager [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Received event network-changed-36b058b4-6728-4495-84e4-6059c1ffed46 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 686.996684] env[61986]: DEBUG nova.compute.manager [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Refreshing instance network info cache due to event network-changed-36b058b4-6728-4495-84e4-6059c1ffed46. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 686.996845] env[61986]: DEBUG oslo_concurrency.lockutils [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] Acquiring lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.996990] env[61986]: DEBUG oslo_concurrency.lockutils [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] Acquired lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.997743] env[61986]: DEBUG nova.network.neutron [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Refreshing network info cache for port 36b058b4-6728-4495-84e4-6059c1ffed46 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 687.024228] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Releasing lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.024663] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.024852] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 687.025367] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-877edb07-256e-411f-b959-8ae277859a6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.035309] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2bbd4c-27f2-4a83-873c-911458fa05fc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.062110] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 35e41d19-8f1c-496d-9712-a1ac3c787c3b could not be found. [ 687.062348] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 687.062532] env[61986]: INFO nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 687.062775] env[61986]: DEBUG oslo.service.loopingcall [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.064945] env[61986]: DEBUG nova.compute.manager [-] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.064945] env[61986]: DEBUG nova.network.neutron [-] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.081671] env[61986]: DEBUG nova.network.neutron [-] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.237116] env[61986]: ERROR nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 687.237116] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.237116] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.237116] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.237116] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.237116] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.237116] env[61986]: ERROR nova.compute.manager raise self.value [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.237116] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.237116] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.237116] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.237604] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.237604] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.237604] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 687.237604] env[61986]: ERROR nova.compute.manager [ 687.237604] env[61986]: Traceback (most recent call last): [ 687.237604] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.237604] env[61986]: listener.cb(fileno) [ 687.237604] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.237604] env[61986]: result = function(*args, **kwargs) [ 687.237604] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.237604] env[61986]: return func(*args, **kwargs) [ 687.237604] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.237604] env[61986]: raise e [ 687.237604] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.237604] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 687.237604] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.237604] env[61986]: created_port_ids = self._update_ports_for_instance( [ 687.237604] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.237604] env[61986]: with excutils.save_and_reraise_exception(): [ 687.237604] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.237604] env[61986]: self.force_reraise() [ 687.237604] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.237604] env[61986]: raise self.value [ 687.237604] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.237604] env[61986]: updated_port = self._update_port( [ 687.237604] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.237604] env[61986]: _ensure_no_port_binding_failure(port) [ 687.237604] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.237604] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.238491] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 687.238491] env[61986]: Removing descriptor: 17 [ 687.265282] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.269029] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.269029] env[61986]: ERROR nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Traceback (most recent call last): [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self.driver.spawn(context, instance, image_meta, [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.269029] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] vm_ref = self.build_virtual_machine(instance, [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] for vif in network_info: [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return self._sync_wrapper(fn, *args, **kwargs) [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self.wait() [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self[:] = self._gt.wait() [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return self._exit_event.wait() [ 687.269421] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] result = hub.switch() [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return self.greenlet.switch() [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] result = function(*args, **kwargs) [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] return func(*args, **kwargs) [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] raise e [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] nwinfo = self.network_api.allocate_for_instance( [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.269782] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] created_port_ids = self._update_ports_for_instance( [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] with excutils.save_and_reraise_exception(): [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] self.force_reraise() [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] raise self.value [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] updated_port = self._update_port( [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] _ensure_no_port_binding_failure(port) [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.270140] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] raise exception.PortBindingFailed(port_id=port['id']) [ 687.270443] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] nova.exception.PortBindingFailed: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. [ 687.270443] env[61986]: ERROR nova.compute.manager [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] [ 687.270443] env[61986]: DEBUG nova.compute.utils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.270836] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.284s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.272311] env[61986]: INFO nova.compute.claims [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.274957] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Build of instance 10568cfc-3634-47d7-96d6-893c4fe01282 was re-scheduled: Binding failed for port ddd2a3b6-c8c7-4cbc-9292-e01f1b78f821, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.275554] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.275799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.275954] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.276123] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.298234] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.298496] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.298648] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.298849] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.299050] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.299174] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.299371] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.299523] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.299691] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.299851] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.300035] env[61986]: DEBUG nova.virt.hardware [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.301271] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b41e20-2f97-4bed-9fdc-c51c6d960a8e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.310193] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4024d5-fecd-46ba-b54b-63dc93c4ee65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.326267] env[61986]: ERROR nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Traceback (most recent call last): [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] yield resources [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self.driver.spawn(context, instance, image_meta, [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] vm_ref = self.build_virtual_machine(instance, [ 687.326267] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] for vif in network_info: [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] return self._sync_wrapper(fn, *args, **kwargs) [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self.wait() [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self[:] = self._gt.wait() [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] return self._exit_event.wait() [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.326652] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] current.throw(*self._exc) [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] result = function(*args, **kwargs) [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] return func(*args, **kwargs) [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] raise e [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] nwinfo = self.network_api.allocate_for_instance( [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] created_port_ids = self._update_ports_for_instance( [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] with excutils.save_and_reraise_exception(): [ 687.327020] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self.force_reraise() [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] raise self.value [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] updated_port = self._update_port( [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] _ensure_no_port_binding_failure(port) [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] raise exception.PortBindingFailed(port_id=port['id']) [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 687.327420] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] [ 687.327420] env[61986]: INFO nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Terminating instance [ 687.328524] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquiring lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.582655] env[61986]: DEBUG nova.network.neutron [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.586739] env[61986]: DEBUG nova.network.neutron [-] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.666687] env[61986]: DEBUG nova.network.neutron [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.799473] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.932080] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.089133] env[61986]: INFO nova.compute.manager [-] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Took 1.03 seconds to deallocate network for instance. [ 688.091422] env[61986]: DEBUG nova.compute.claims [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 688.091591] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.172686] env[61986]: DEBUG oslo_concurrency.lockutils [req-6d915c58-500c-4569-b171-c94da2b98655 req-5b0da78c-64f8-410f-aa1e-612c4e95913a service nova] Releasing lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.173122] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquired lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.173306] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.435678] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-10568cfc-3634-47d7-96d6-893c4fe01282" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.435945] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 688.436214] env[61986]: DEBUG nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.436447] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 688.456663] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.704140] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.721484] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45b5abd-b43d-456e-a451-087e3576f86d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.730149] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a42dc-fd37-4a05-aa87-871ee101a2a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.759835] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4430c2cb-6978-476c-a554-f2324ae3d792 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.771806] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d1073b-46e4-4bb8-b18b-4dbf4a5f7d8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.779761] env[61986]: DEBUG nova.compute.provider_tree [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.784380] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.961392] env[61986]: DEBUG nova.network.neutron [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.065223] env[61986]: DEBUG nova.compute.manager [req-c77e2ae2-68d2-4061-8571-b39e9fbab392 req-15ba0c7f-e5c3-434c-a883-1078a314b0b1 service nova] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Received event network-vif-deleted-36b058b4-6728-4495-84e4-6059c1ffed46 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 689.282660] env[61986]: DEBUG nova.scheduler.client.report [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.286853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Releasing lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.287259] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 689.287451] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 689.287721] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a0c1c77-7753-4786-9386-79fc03304cfb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.297019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041c9ab8-3832-4f6d-be03-2005490ac49c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.324402] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4fa110e0-7235-4d62-a68b-c727b4a79908 could not be found. [ 689.324623] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 689.324810] env[61986]: INFO nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Took 0.04 seconds to destroy the instance on the hypervisor. [ 689.325047] env[61986]: DEBUG oslo.service.loopingcall [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.325255] env[61986]: DEBUG nova.compute.manager [-] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.325350] env[61986]: DEBUG nova.network.neutron [-] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 689.343136] env[61986]: DEBUG nova.network.neutron [-] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.464305] env[61986]: INFO nova.compute.manager [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 10568cfc-3634-47d7-96d6-893c4fe01282] Took 1.03 seconds to deallocate network for instance. [ 689.793599] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.794220] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.796902] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.602s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.798390] env[61986]: INFO nova.compute.claims [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.845744] env[61986]: DEBUG nova.network.neutron [-] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.140960] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "8438d2a7-975c-4d1f-8906-f960bfadfe39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.141251] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.190448] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 690.190766] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 690.305037] env[61986]: DEBUG nova.compute.utils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.307057] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.307247] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.350221] env[61986]: INFO nova.compute.manager [-] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Took 1.02 seconds to deallocate network for instance. [ 690.351434] env[61986]: DEBUG nova.compute.claims [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 690.351846] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.353757] env[61986]: DEBUG nova.policy [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd02f6597b238409cb0b87adcbdf96626', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '280e38fa45ac4c80a47617fbd5d4e4c6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 690.498751] env[61986]: INFO nova.scheduler.client.report [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted allocations for instance 10568cfc-3634-47d7-96d6-893c4fe01282 [ 690.641858] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Successfully created port: f6d8e712-dc02-4bea-85d3-8166d0a31740 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.702646] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 690.704643] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 690.704880] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Rebuilding the list of instances to heal {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 690.809300] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.008293] env[61986]: DEBUG oslo_concurrency.lockutils [None req-eea4b534-91d4-4574-a82a-d6c41be39d3e tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "10568cfc-3634-47d7-96d6-893c4fe01282" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.834s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.718250] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 691.718490] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 691.718578] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 691.718705] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 691.718821] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 691.718937] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 691.719066] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Didn't find any instances for network info cache update. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 691.722384] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.728374] env[61986]: DEBUG nova.compute.manager [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Received event network-changed-f6d8e712-dc02-4bea-85d3-8166d0a31740 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 691.728549] env[61986]: DEBUG nova.compute.manager [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Refreshing instance network info cache due to event network-changed-f6d8e712-dc02-4bea-85d3-8166d0a31740. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 691.728759] env[61986]: DEBUG oslo_concurrency.lockutils [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] Acquiring lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.728896] env[61986]: DEBUG oslo_concurrency.lockutils [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] Acquired lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.729061] env[61986]: DEBUG nova.network.neutron [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Refreshing network info cache for port f6d8e712-dc02-4bea-85d3-8166d0a31740 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 691.730325] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.730682] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.731328] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.731504] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.731648] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.731792] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.731915] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 691.732064] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 691.835278] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa450026-64bb-4bc5-b45a-3bb3ef869df2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.843218] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e8462d-971f-4e56-9829-dca5f878ace7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.872672] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdab84d2-3bd7-4b34-808f-236d4e9e0774 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.879992] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d11225-a559-49ca-ba7f-9043a8f6355b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.893171] env[61986]: DEBUG nova.compute.provider_tree [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.945631] env[61986]: ERROR nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 691.945631] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.945631] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.945631] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.945631] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.945631] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.945631] env[61986]: ERROR nova.compute.manager raise self.value [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.945631] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.945631] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.945631] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.946318] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.946318] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.946318] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 691.946318] env[61986]: ERROR nova.compute.manager [ 691.946318] env[61986]: Traceback (most recent call last): [ 691.946318] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.946318] env[61986]: listener.cb(fileno) [ 691.946318] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.946318] env[61986]: result = function(*args, **kwargs) [ 691.946318] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.946318] env[61986]: return func(*args, **kwargs) [ 691.946318] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.946318] env[61986]: raise e [ 691.946318] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.946318] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 691.946318] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.946318] env[61986]: created_port_ids = self._update_ports_for_instance( [ 691.946318] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.946318] env[61986]: with excutils.save_and_reraise_exception(): [ 691.946318] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.946318] env[61986]: self.force_reraise() [ 691.946318] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.946318] env[61986]: raise self.value [ 691.946318] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.946318] env[61986]: updated_port = self._update_port( [ 691.946318] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.946318] env[61986]: _ensure_no_port_binding_failure(port) [ 691.946318] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.946318] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.947381] env[61986]: nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 691.947381] env[61986]: Removing descriptor: 17 [ 692.228935] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.235616] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.252465] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.255819] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.256046] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.256204] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.256385] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.256568] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.256723] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.256924] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.257087] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.257250] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.257405] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.257565] env[61986]: DEBUG nova.virt.hardware [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.258759] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998fd6c5-25a0-4e3f-94d9-4299015d6687 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.261770] env[61986]: DEBUG nova.network.neutron [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.268904] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a4917f-0ea2-4107-91f4-f02a1d531e81 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.281876] env[61986]: ERROR nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Traceback (most recent call last): [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] yield resources [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self.driver.spawn(context, instance, image_meta, [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] vm_ref = self.build_virtual_machine(instance, [ 692.281876] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] for vif in network_info: [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] return self._sync_wrapper(fn, *args, **kwargs) [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self.wait() [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self[:] = self._gt.wait() [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] return self._exit_event.wait() [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.282221] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] current.throw(*self._exc) [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] result = function(*args, **kwargs) [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] return func(*args, **kwargs) [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] raise e [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] nwinfo = self.network_api.allocate_for_instance( [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] created_port_ids = self._update_ports_for_instance( [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] with excutils.save_and_reraise_exception(): [ 692.282569] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self.force_reraise() [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] raise self.value [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] updated_port = self._update_port( [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] _ensure_no_port_binding_failure(port) [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] raise exception.PortBindingFailed(port_id=port['id']) [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 692.282919] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] [ 692.282919] env[61986]: INFO nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Terminating instance [ 692.285744] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquiring lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.349836] env[61986]: DEBUG nova.network.neutron [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.396520] env[61986]: DEBUG nova.scheduler.client.report [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.427442] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.427673] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.852779] env[61986]: DEBUG oslo_concurrency.lockutils [req-412e5997-c9c3-479a-882c-9e0c0e679bda req-0191ddcf-8266-40fe-bee1-5459b514e5bf service nova] Releasing lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.853060] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquired lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.853343] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.901110] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.104s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.901631] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.904648] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.548s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.906067] env[61986]: INFO nova.compute.claims [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.371553] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.412881] env[61986]: DEBUG nova.compute.utils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.414313] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.414437] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.470491] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.477557] env[61986]: DEBUG nova.policy [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b3e2ba7061649cfaf5d13ed7f2ba129', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '214f8e64bea7498198ab9feff6b2d0ef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 693.487449] env[61986]: DEBUG nova.compute.manager [req-4ddb15d6-cf0f-45cc-addc-bf690b00b726 req-900de3e7-560f-48ef-9ccc-190a56580af5 service nova] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Received event network-vif-deleted-f6d8e712-dc02-4bea-85d3-8166d0a31740 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 693.793200] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Successfully created port: b1498699-69a3-44e4-bd22-edfd4f7f3633 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.917961] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.977058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Releasing lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.977058] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 693.977058] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 693.978795] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-182f5db4-deba-4b29-a35a-36ed462b3246 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.987647] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e21a25-ff6d-4728-9936-2f1a810bc600 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.016309] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b632a5d1-62b1-44b1-a82a-b1e761e0c4b8 could not be found. [ 694.016309] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 694.016309] env[61986]: INFO nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.016761] env[61986]: DEBUG oslo.service.loopingcall [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.017106] env[61986]: DEBUG nova.compute.manager [-] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.017348] env[61986]: DEBUG nova.network.neutron [-] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 694.037190] env[61986]: DEBUG nova.network.neutron [-] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.342571] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d416f7b4-bea3-489a-a062-0eb8b439dfcf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.352201] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e1f12c-910b-4266-bfef-4747e4ff10d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.381318] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9d2a1b-7cf5-419d-a833-b30fe1491878 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.388196] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9e7b44-6427-4ee4-90fb-6615a326a540 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.400916] env[61986]: DEBUG nova.compute.provider_tree [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.540785] env[61986]: DEBUG nova.network.neutron [-] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.812538] env[61986]: DEBUG nova.compute.manager [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Received event network-changed-b1498699-69a3-44e4-bd22-edfd4f7f3633 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 694.812862] env[61986]: DEBUG nova.compute.manager [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Refreshing instance network info cache due to event network-changed-b1498699-69a3-44e4-bd22-edfd4f7f3633. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 694.813131] env[61986]: DEBUG oslo_concurrency.lockutils [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] Acquiring lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.813311] env[61986]: DEBUG oslo_concurrency.lockutils [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] Acquired lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.813508] env[61986]: DEBUG nova.network.neutron [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Refreshing network info cache for port b1498699-69a3-44e4-bd22-edfd4f7f3633 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 694.904700] env[61986]: DEBUG nova.scheduler.client.report [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.929200] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.957132] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.957438] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.957639] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.957831] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.957975] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.958130] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.958336] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.958485] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.958643] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.958797] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.958963] env[61986]: DEBUG nova.virt.hardware [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.960070] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38405e70-78aa-47eb-bdc4-96ddab417501 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.967935] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abcd957-78c5-4fdb-adcf-897d880384ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.989680] env[61986]: ERROR nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 694.989680] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.989680] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.989680] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.989680] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.989680] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.989680] env[61986]: ERROR nova.compute.manager raise self.value [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.989680] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.989680] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.989680] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.990161] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.990161] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.990161] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 694.990161] env[61986]: ERROR nova.compute.manager [ 694.990161] env[61986]: Traceback (most recent call last): [ 694.990161] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.990161] env[61986]: listener.cb(fileno) [ 694.990161] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.990161] env[61986]: result = function(*args, **kwargs) [ 694.990161] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.990161] env[61986]: return func(*args, **kwargs) [ 694.990161] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.990161] env[61986]: raise e [ 694.990161] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.990161] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 694.990161] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.990161] env[61986]: created_port_ids = self._update_ports_for_instance( [ 694.990161] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.990161] env[61986]: with excutils.save_and_reraise_exception(): [ 694.990161] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.990161] env[61986]: self.force_reraise() [ 694.990161] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.990161] env[61986]: raise self.value [ 694.990161] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.990161] env[61986]: updated_port = self._update_port( [ 694.990161] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.990161] env[61986]: _ensure_no_port_binding_failure(port) [ 694.990161] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.990161] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.991031] env[61986]: nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 694.991031] env[61986]: Removing descriptor: 14 [ 694.991031] env[61986]: ERROR nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Traceback (most recent call last): [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] yield resources [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self.driver.spawn(context, instance, image_meta, [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.991031] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] vm_ref = self.build_virtual_machine(instance, [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] for vif in network_info: [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return self._sync_wrapper(fn, *args, **kwargs) [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self.wait() [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self[:] = self._gt.wait() [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return self._exit_event.wait() [ 694.991356] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] result = hub.switch() [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return self.greenlet.switch() [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] result = function(*args, **kwargs) [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return func(*args, **kwargs) [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] raise e [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] nwinfo = self.network_api.allocate_for_instance( [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.991710] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] created_port_ids = self._update_ports_for_instance( [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] with excutils.save_and_reraise_exception(): [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self.force_reraise() [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] raise self.value [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] updated_port = self._update_port( [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] _ensure_no_port_binding_failure(port) [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.992143] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] raise exception.PortBindingFailed(port_id=port['id']) [ 694.992413] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 694.992413] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] [ 694.992413] env[61986]: INFO nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Terminating instance [ 694.992846] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquiring lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.043523] env[61986]: INFO nova.compute.manager [-] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Took 1.03 seconds to deallocate network for instance. [ 695.045836] env[61986]: DEBUG nova.compute.claims [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 695.046149] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.337306] env[61986]: DEBUG nova.network.neutron [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.410375] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.410881] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.413437] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.748s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.414765] env[61986]: INFO nova.compute.claims [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.460972] env[61986]: DEBUG nova.network.neutron [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.920046] env[61986]: DEBUG nova.compute.utils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.922951] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.923137] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.964252] env[61986]: DEBUG oslo_concurrency.lockutils [req-5a473e96-6d0e-41a8-ad67-1a621ac47307 req-b2832ec4-5bdd-4f5e-877d-4ab99a3095ff service nova] Releasing lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.964666] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquired lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.964851] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.980753] env[61986]: DEBUG nova.policy [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c8822830e4944d2ad65cc22a32e24fd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec92f06a60c04d4a9aec4f7d245185aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 696.423842] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.462484] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Successfully created port: 71e43a41-e010-4585-ad34-d2e52d8730b7 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.489413] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.572879] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.831120] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff3f9ae-fcfb-485f-9c18-487c6872ed3e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.840755] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bff59a3-5621-4d3a-a323-8e3f38d4e4e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.845279] env[61986]: DEBUG nova.compute.manager [req-d931fd24-363d-459d-af27-5e3edea9d42f req-91985af6-c09c-4ec7-9df8-59bbc63a8fa6 service nova] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Received event network-vif-deleted-b1498699-69a3-44e4-bd22-edfd4f7f3633 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.875484] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64033d27-8be7-4b2e-b4cd-e96a3936e6be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.882911] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b0a9de-c535-42c3-9688-4bdecdaf7046 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.898401] env[61986]: DEBUG nova.compute.provider_tree [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.075722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Releasing lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.076273] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.076472] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.076788] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e5cd422-8c35-4280-8535-67d923c58847 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.085690] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9974f1f0-219f-401d-a248-b4c3a0c68b80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.109941] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cbdf7e34-57f9-4456-8cc7-f09c15e0f965 could not be found. [ 697.110195] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 697.110593] env[61986]: INFO nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Took 0.03 seconds to destroy the instance on the hypervisor. [ 697.110593] env[61986]: DEBUG oslo.service.loopingcall [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.110821] env[61986]: DEBUG nova.compute.manager [-] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.110921] env[61986]: DEBUG nova.network.neutron [-] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.129115] env[61986]: DEBUG nova.network.neutron [-] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.404019] env[61986]: DEBUG nova.scheduler.client.report [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.435113] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 697.461434] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.463016] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.463016] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.463016] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.463016] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.463016] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.463207] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.463207] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.463207] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.463554] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.463855] env[61986]: DEBUG nova.virt.hardware [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.464784] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee26dc4-ea54-4f62-ac77-86b81fff90f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.473625] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09bf653-2b37-4b79-8f7f-d9b715e0f997 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.634476] env[61986]: DEBUG nova.network.neutron [-] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.853294] env[61986]: ERROR nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 697.853294] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.853294] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.853294] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.853294] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.853294] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.853294] env[61986]: ERROR nova.compute.manager raise self.value [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.853294] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 697.853294] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.853294] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 697.853707] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.853707] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 697.853707] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 697.853707] env[61986]: ERROR nova.compute.manager [ 697.853707] env[61986]: Traceback (most recent call last): [ 697.853707] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 697.853707] env[61986]: listener.cb(fileno) [ 697.853707] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.853707] env[61986]: result = function(*args, **kwargs) [ 697.853707] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.853707] env[61986]: return func(*args, **kwargs) [ 697.853707] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.853707] env[61986]: raise e [ 697.853707] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.853707] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 697.853707] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.853707] env[61986]: created_port_ids = self._update_ports_for_instance( [ 697.853707] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.853707] env[61986]: with excutils.save_and_reraise_exception(): [ 697.853707] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.853707] env[61986]: self.force_reraise() [ 697.853707] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.853707] env[61986]: raise self.value [ 697.853707] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.853707] env[61986]: updated_port = self._update_port( [ 697.853707] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.853707] env[61986]: _ensure_no_port_binding_failure(port) [ 697.853707] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.853707] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 697.854357] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 697.854357] env[61986]: Removing descriptor: 14 [ 697.854357] env[61986]: ERROR nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Traceback (most recent call last): [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] yield resources [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self.driver.spawn(context, instance, image_meta, [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.854357] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] vm_ref = self.build_virtual_machine(instance, [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] for vif in network_info: [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return self._sync_wrapper(fn, *args, **kwargs) [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self.wait() [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self[:] = self._gt.wait() [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return self._exit_event.wait() [ 697.854621] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] result = hub.switch() [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return self.greenlet.switch() [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] result = function(*args, **kwargs) [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return func(*args, **kwargs) [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] raise e [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] nwinfo = self.network_api.allocate_for_instance( [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.854901] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] created_port_ids = self._update_ports_for_instance( [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] with excutils.save_and_reraise_exception(): [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self.force_reraise() [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] raise self.value [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] updated_port = self._update_port( [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] _ensure_no_port_binding_failure(port) [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.855199] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] raise exception.PortBindingFailed(port_id=port['id']) [ 697.855460] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 697.855460] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] [ 697.855460] env[61986]: INFO nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Terminating instance [ 697.856660] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquiring lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.856871] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquired lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.857062] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.906324] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.906840] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.909530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.028s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.141047] env[61986]: INFO nova.compute.manager [-] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Took 1.03 seconds to deallocate network for instance. [ 698.142732] env[61986]: DEBUG nova.compute.claims [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 698.142907] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.384845] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.416014] env[61986]: DEBUG nova.compute.utils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.420353] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.420504] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 698.442426] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.460231] env[61986]: DEBUG nova.policy [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d6bc369cf4a45b7b15ef85c8803cb25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8a569db0be340a5999a985f41853578', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 698.731478] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Successfully created port: 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.824357] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de9dcd7-20f2-4db9-91bf-cbca72aa694b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.832913] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a13c247-5908-49f8-8926-6ad627481e95 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.867382] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313575e9-573f-4acc-ac35-3be351034dd9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.875236] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c42fbc-fd26-4e05-b872-5a353455f718 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.888838] env[61986]: DEBUG nova.compute.provider_tree [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.911351] env[61986]: DEBUG nova.compute.manager [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Received event network-changed-71e43a41-e010-4585-ad34-d2e52d8730b7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.911351] env[61986]: DEBUG nova.compute.manager [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Refreshing instance network info cache due to event network-changed-71e43a41-e010-4585-ad34-d2e52d8730b7. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 698.911351] env[61986]: DEBUG oslo_concurrency.lockutils [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] Acquiring lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.920661] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.946876] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Releasing lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.947317] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 698.947496] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 698.947791] env[61986]: DEBUG oslo_concurrency.lockutils [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] Acquired lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.947950] env[61986]: DEBUG nova.network.neutron [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Refreshing network info cache for port 71e43a41-e010-4585-ad34-d2e52d8730b7 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 698.949045] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e55b6dee-7c9e-41ba-ab00-a9811d6c0df6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.959421] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3123711-6154-47fa-a9b5-f29884f06c1d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.982083] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bf115dff-5440-41ba-9689-6337f63dedeb could not be found. [ 698.982312] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 698.982491] env[61986]: INFO nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 698.982735] env[61986]: DEBUG oslo.service.loopingcall [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.982936] env[61986]: DEBUG nova.compute.manager [-] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.983118] env[61986]: DEBUG nova.network.neutron [-] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.999652] env[61986]: DEBUG nova.network.neutron [-] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.391455] env[61986]: DEBUG nova.scheduler.client.report [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.446679] env[61986]: DEBUG nova.compute.manager [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Received event network-changed-0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 699.447033] env[61986]: DEBUG nova.compute.manager [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Refreshing instance network info cache due to event network-changed-0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 699.447172] env[61986]: DEBUG oslo_concurrency.lockutils [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] Acquiring lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.447260] env[61986]: DEBUG oslo_concurrency.lockutils [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] Acquired lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.447422] env[61986]: DEBUG nova.network.neutron [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Refreshing network info cache for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 699.470966] env[61986]: DEBUG nova.network.neutron [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.501760] env[61986]: DEBUG nova.network.neutron [-] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.539689] env[61986]: DEBUG nova.network.neutron [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.581573] env[61986]: ERROR nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 699.581573] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.581573] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.581573] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.581573] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.581573] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.581573] env[61986]: ERROR nova.compute.manager raise self.value [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.581573] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.581573] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.581573] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.581957] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.581957] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.581957] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 699.581957] env[61986]: ERROR nova.compute.manager [ 699.581957] env[61986]: Traceback (most recent call last): [ 699.581957] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.581957] env[61986]: listener.cb(fileno) [ 699.581957] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.581957] env[61986]: result = function(*args, **kwargs) [ 699.581957] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.581957] env[61986]: return func(*args, **kwargs) [ 699.581957] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.581957] env[61986]: raise e [ 699.581957] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.581957] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 699.581957] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.581957] env[61986]: created_port_ids = self._update_ports_for_instance( [ 699.581957] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.581957] env[61986]: with excutils.save_and_reraise_exception(): [ 699.581957] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.581957] env[61986]: self.force_reraise() [ 699.581957] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.581957] env[61986]: raise self.value [ 699.581957] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.581957] env[61986]: updated_port = self._update_port( [ 699.581957] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.581957] env[61986]: _ensure_no_port_binding_failure(port) [ 699.581957] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.581957] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.583194] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 699.583194] env[61986]: Removing descriptor: 17 [ 699.896750] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.897556] env[61986]: ERROR nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Traceback (most recent call last): [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self.driver.spawn(context, instance, image_meta, [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] vm_ref = self.build_virtual_machine(instance, [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.897556] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] for vif in network_info: [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return self._sync_wrapper(fn, *args, **kwargs) [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self.wait() [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self[:] = self._gt.wait() [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return self._exit_event.wait() [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] result = hub.switch() [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.897844] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return self.greenlet.switch() [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] result = function(*args, **kwargs) [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] return func(*args, **kwargs) [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] raise e [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] nwinfo = self.network_api.allocate_for_instance( [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] created_port_ids = self._update_ports_for_instance( [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] with excutils.save_and_reraise_exception(): [ 699.898147] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] self.force_reraise() [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] raise self.value [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] updated_port = self._update_port( [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] _ensure_no_port_binding_failure(port) [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] raise exception.PortBindingFailed(port_id=port['id']) [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] nova.exception.PortBindingFailed: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. [ 699.898491] env[61986]: ERROR nova.compute.manager [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] [ 699.898775] env[61986]: DEBUG nova.compute.utils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.899605] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.975s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.903239] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Build of instance 236ecde7-1250-4ce1-837c-8266b95dd513 was re-scheduled: Binding failed for port 255384ec-de5e-4e9e-8371-c22f2e0f32fa, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.903314] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.903530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.903674] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.903869] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.934093] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.960709] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.960954] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.961569] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.961569] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.961569] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.961765] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.961811] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.962417] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.962417] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.962417] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.962417] env[61986]: DEBUG nova.virt.hardware [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.963281] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c22188-4c65-4bc5-bac1-e75fb1207db6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.971347] env[61986]: DEBUG nova.network.neutron [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.974196] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96240a72-7b91-4db6-8a51-8f3e17ed18e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.990018] env[61986]: ERROR nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Traceback (most recent call last): [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] yield resources [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self.driver.spawn(context, instance, image_meta, [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] vm_ref = self.build_virtual_machine(instance, [ 699.990018] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] for vif in network_info: [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] return self._sync_wrapper(fn, *args, **kwargs) [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self.wait() [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self[:] = self._gt.wait() [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] return self._exit_event.wait() [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.990299] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] current.throw(*self._exc) [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] result = function(*args, **kwargs) [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] return func(*args, **kwargs) [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] raise e [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] nwinfo = self.network_api.allocate_for_instance( [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] created_port_ids = self._update_ports_for_instance( [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] with excutils.save_and_reraise_exception(): [ 699.990606] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self.force_reraise() [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] raise self.value [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] updated_port = self._update_port( [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] _ensure_no_port_binding_failure(port) [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] raise exception.PortBindingFailed(port_id=port['id']) [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 699.990966] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] [ 699.990966] env[61986]: INFO nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Terminating instance [ 699.995823] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquiring lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.004385] env[61986]: INFO nova.compute.manager [-] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Took 1.02 seconds to deallocate network for instance. [ 700.006643] env[61986]: DEBUG nova.compute.claims [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 700.006840] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.041907] env[61986]: DEBUG oslo_concurrency.lockutils [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] Releasing lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.042164] env[61986]: DEBUG nova.compute.manager [req-1f038a4f-049b-4b88-a524-439224718380 req-f0e44df2-ab98-4927-882c-77c1eff5201a service nova] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Received event network-vif-deleted-71e43a41-e010-4585-ad34-d2e52d8730b7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 700.053025] env[61986]: DEBUG nova.network.neutron [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.426044] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.503792] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.556180] env[61986]: DEBUG oslo_concurrency.lockutils [req-7ec7eb00-47d1-4361-910e-087da1276334 req-f87edad5-1b43-4344-9c24-07d3ddd97a68 service nova] Releasing lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.556580] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquired lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.556761] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 700.739601] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e3e89f-4941-4b51-8bb1-ee438850442c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.746838] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905c788b-d9c3-4654-ac55-f892b3dbf395 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.779151] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182615a9-5a0b-404f-b9b7-3beed88005fd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.786356] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b710eb-a9ee-4c9a-b642-df78cf035e27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.799721] env[61986]: DEBUG nova.compute.provider_tree [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.006818] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-236ecde7-1250-4ce1-837c-8266b95dd513" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.007089] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 701.007278] env[61986]: DEBUG nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.007445] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 701.022371] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.073935] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.141368] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.302282] env[61986]: DEBUG nova.scheduler.client.report [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.470208] env[61986]: DEBUG nova.compute.manager [req-924da942-47e6-458a-a116-db98d2ef27a7 req-c2e04a35-c900-43e1-90a2-fed753235a26 service nova] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Received event network-vif-deleted-0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.524560] env[61986]: DEBUG nova.network.neutron [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.646628] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Releasing lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.646628] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.646628] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 701.646628] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63f7f450-3ca4-44fd-853c-82537bfc1b97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.654474] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0d885b-c75c-4989-aa58-e6323922f8c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.675522] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f8b9aca-420f-4314-a466-0d6f3d191377 could not be found. [ 701.675679] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 701.675879] env[61986]: INFO nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Took 0.03 seconds to destroy the instance on the hypervisor. [ 701.676129] env[61986]: DEBUG oslo.service.loopingcall [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.676360] env[61986]: DEBUG nova.compute.manager [-] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.676452] env[61986]: DEBUG nova.network.neutron [-] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 701.692175] env[61986]: DEBUG nova.network.neutron [-] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.807150] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.807784] env[61986]: ERROR nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] Traceback (most recent call last): [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self.driver.spawn(context, instance, image_meta, [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] vm_ref = self.build_virtual_machine(instance, [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.807784] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] for vif in network_info: [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return self._sync_wrapper(fn, *args, **kwargs) [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self.wait() [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self[:] = self._gt.wait() [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return self._exit_event.wait() [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] result = hub.switch() [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.808166] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return self.greenlet.switch() [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] result = function(*args, **kwargs) [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] return func(*args, **kwargs) [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] raise e [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] nwinfo = self.network_api.allocate_for_instance( [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] created_port_ids = self._update_ports_for_instance( [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] with excutils.save_and_reraise_exception(): [ 701.808689] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] self.force_reraise() [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] raise self.value [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] updated_port = self._update_port( [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] _ensure_no_port_binding_failure(port) [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] raise exception.PortBindingFailed(port_id=port['id']) [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] nova.exception.PortBindingFailed: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. [ 701.809635] env[61986]: ERROR nova.compute.manager [instance: f70adc74-3b24-434a-829a-13274bae40b1] [ 701.809968] env[61986]: DEBUG nova.compute.utils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.809968] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.874s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.811310] env[61986]: INFO nova.compute.claims [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.814103] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Build of instance f70adc74-3b24-434a-829a-13274bae40b1 was re-scheduled: Binding failed for port 0a254c6b-1159-4515-ad39-1bd51a9624aa, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.814558] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.814782] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquiring lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.815376] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Acquired lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.815376] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.027089] env[61986]: INFO nova.compute.manager [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 236ecde7-1250-4ce1-837c-8266b95dd513] Took 1.02 seconds to deallocate network for instance. [ 702.195196] env[61986]: DEBUG nova.network.neutron [-] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.336194] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.406280] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.697446] env[61986]: INFO nova.compute.manager [-] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Took 1.02 seconds to deallocate network for instance. [ 702.700337] env[61986]: DEBUG nova.compute.claims [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 702.700581] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.908583] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Releasing lock "refresh_cache-f70adc74-3b24-434a-829a-13274bae40b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.908814] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.909033] env[61986]: DEBUG nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.909405] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 702.927172] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.055665] env[61986]: INFO nova.scheduler.client.report [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocations for instance 236ecde7-1250-4ce1-837c-8266b95dd513 [ 703.241295] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07de1834-14ee-481d-a7c3-ecd5725ebbfb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.248721] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d907497-daa1-45b2-b935-1b5cca756ef8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.277747] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cce9b27-9461-4b28-a824-84fd0ab16ed5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.284964] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b854621e-35a1-4898-a71f-a2fdba18b80e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.298765] env[61986]: DEBUG nova.compute.provider_tree [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.430069] env[61986]: DEBUG nova.network.neutron [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.569569] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a342d292-5341-42c3-9aa2-cca2bb464a5f tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "236ecde7-1250-4ce1-837c-8266b95dd513" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.123s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.802023] env[61986]: DEBUG nova.scheduler.client.report [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.934365] env[61986]: INFO nova.compute.manager [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] [instance: f70adc74-3b24-434a-829a-13274bae40b1] Took 1.02 seconds to deallocate network for instance. [ 704.072533] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 704.306703] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.307222] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.310592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.219s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.358376] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.358376] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.595512] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.816049] env[61986]: DEBUG nova.compute.utils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.820242] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.820344] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.873227] env[61986]: DEBUG nova.policy [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c508fb3095b4d5ebe65ec78017fdd28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a353820fb5cf4b8aa6ba875dc79c8043', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 704.961338] env[61986]: INFO nova.scheduler.client.report [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Deleted allocations for instance f70adc74-3b24-434a-829a-13274bae40b1 [ 705.181375] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb54bad9-5850-40b7-a15f-0d475db0350c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.195282] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9a904d-819b-4172-b682-308306a2a3ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.199561] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Successfully created port: 06a228ab-e02b-4472-a877-22ee30cb8d8e {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.233625] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b7d2d2-7759-468e-8197-e80a6b213e02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.242243] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4819723-4b25-40c6-b496-58b6e22e3128 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.260482] env[61986]: DEBUG nova.compute.provider_tree [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.321332] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.471067] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bab2ba5-da1c-42f7-838e-d172ac9744b2 tempest-ServerMetadataTestJSON-1597644333 tempest-ServerMetadataTestJSON-1597644333-project-member] Lock "f70adc74-3b24-434a-829a-13274bae40b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.847s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.766018] env[61986]: DEBUG nova.scheduler.client.report [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.851734] env[61986]: DEBUG nova.compute.manager [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Received event network-changed-06a228ab-e02b-4472-a877-22ee30cb8d8e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.852579] env[61986]: DEBUG nova.compute.manager [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Refreshing instance network info cache due to event network-changed-06a228ab-e02b-4472-a877-22ee30cb8d8e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 705.852698] env[61986]: DEBUG oslo_concurrency.lockutils [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] Acquiring lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.852840] env[61986]: DEBUG oslo_concurrency.lockutils [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] Acquired lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.853242] env[61986]: DEBUG nova.network.neutron [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Refreshing network info cache for port 06a228ab-e02b-4472-a877-22ee30cb8d8e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 705.973995] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.159543] env[61986]: ERROR nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 706.159543] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.159543] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.159543] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.159543] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.159543] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.159543] env[61986]: ERROR nova.compute.manager raise self.value [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.159543] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.159543] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.159543] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.160190] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.160190] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.160190] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 706.160190] env[61986]: ERROR nova.compute.manager [ 706.160190] env[61986]: Traceback (most recent call last): [ 706.160190] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.160190] env[61986]: listener.cb(fileno) [ 706.160190] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.160190] env[61986]: result = function(*args, **kwargs) [ 706.160190] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.160190] env[61986]: return func(*args, **kwargs) [ 706.160190] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.160190] env[61986]: raise e [ 706.160190] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.160190] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 706.160190] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.160190] env[61986]: created_port_ids = self._update_ports_for_instance( [ 706.160190] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.160190] env[61986]: with excutils.save_and_reraise_exception(): [ 706.160190] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.160190] env[61986]: self.force_reraise() [ 706.160190] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.160190] env[61986]: raise self.value [ 706.160190] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.160190] env[61986]: updated_port = self._update_port( [ 706.160190] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.160190] env[61986]: _ensure_no_port_binding_failure(port) [ 706.160190] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.160190] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.161224] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 706.161224] env[61986]: Removing descriptor: 17 [ 706.269929] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.271550] env[61986]: ERROR nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Traceback (most recent call last): [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self.driver.spawn(context, instance, image_meta, [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] vm_ref = self.build_virtual_machine(instance, [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.271550] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] for vif in network_info: [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] return self._sync_wrapper(fn, *args, **kwargs) [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self.wait() [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self[:] = self._gt.wait() [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] return self._exit_event.wait() [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] current.throw(*self._exc) [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.271881] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] result = function(*args, **kwargs) [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] return func(*args, **kwargs) [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] raise e [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] nwinfo = self.network_api.allocate_for_instance( [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] created_port_ids = self._update_ports_for_instance( [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] with excutils.save_and_reraise_exception(): [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] self.force_reraise() [ 706.272184] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] raise self.value [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] updated_port = self._update_port( [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] _ensure_no_port_binding_failure(port) [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] raise exception.PortBindingFailed(port_id=port['id']) [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] nova.exception.PortBindingFailed: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. [ 706.272470] env[61986]: ERROR nova.compute.manager [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] [ 706.272470] env[61986]: DEBUG nova.compute.utils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.274133] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.922s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.277690] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Build of instance 35e41d19-8f1c-496d-9712-a1ac3c787c3b was re-scheduled: Binding failed for port 67d7b46b-c2e9-4979-9ce3-404a43df9063, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.278735] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.279119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquiring lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.279333] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Acquired lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.279530] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.331775] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.359018] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.359018] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.359018] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.359265] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.359265] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.359265] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.359265] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.359265] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.359726] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.360026] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.360331] env[61986]: DEBUG nova.virt.hardware [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.364223] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbbb195-486b-4aca-acce-4f58b7f6b313 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.374060] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db9927a-71ac-480b-8121-9846de8fc837 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.389080] env[61986]: ERROR nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Traceback (most recent call last): [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] yield resources [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self.driver.spawn(context, instance, image_meta, [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] vm_ref = self.build_virtual_machine(instance, [ 706.389080] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] for vif in network_info: [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] return self._sync_wrapper(fn, *args, **kwargs) [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self.wait() [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self[:] = self._gt.wait() [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] return self._exit_event.wait() [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.389416] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] current.throw(*self._exc) [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] result = function(*args, **kwargs) [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] return func(*args, **kwargs) [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] raise e [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] nwinfo = self.network_api.allocate_for_instance( [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] created_port_ids = self._update_ports_for_instance( [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] with excutils.save_and_reraise_exception(): [ 706.389740] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self.force_reraise() [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] raise self.value [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] updated_port = self._update_port( [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] _ensure_no_port_binding_failure(port) [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] raise exception.PortBindingFailed(port_id=port['id']) [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 706.390100] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] [ 706.393021] env[61986]: INFO nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Terminating instance [ 706.393350] env[61986]: DEBUG nova.network.neutron [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.395486] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquiring lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.485967] env[61986]: DEBUG nova.network.neutron [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.510217] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.803313] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.891158] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.991113] env[61986]: DEBUG oslo_concurrency.lockutils [req-e83d8631-e068-4821-a0aa-6135c99cee27 req-615a05b2-676d-42ad-8677-edc72bc296d1 service nova] Releasing lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.992560] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquired lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.992752] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.156916] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ad36c8-41dd-494d-8b01-e38185cb06b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.164649] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c274c1ac-66b8-495e-b11f-d1539072da9a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.195182] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0853e9-3688-4dec-92d0-dd8f18090ee7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.201456] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62aab7f1-56b0-4c43-ab21-a53b396990e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.215604] env[61986]: DEBUG nova.compute.provider_tree [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.395827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Releasing lock "refresh_cache-35e41d19-8f1c-496d-9712-a1ac3c787c3b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.396244] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 707.396326] env[61986]: DEBUG nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.396483] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.411477] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.521699] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.610834] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.718446] env[61986]: DEBUG nova.scheduler.client.report [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.908236] env[61986]: DEBUG nova.compute.manager [req-3dfd38d5-d6e9-4536-803a-a2b5c34a45c6 req-1176a905-f300-4071-bf7e-5239c4dfb976 service nova] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Received event network-vif-deleted-06a228ab-e02b-4472-a877-22ee30cb8d8e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.913333] env[61986]: DEBUG nova.network.neutron [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.113648] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Releasing lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.113753] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.114126] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 708.114486] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dedbf8df-50ea-4825-b42e-1182770c6a8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.123744] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0e64e5-f298-44b7-91c3-342db0347c82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.146180] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a2c58c0a-785b-42a2-9960-0b3e4ff86d96 could not be found. [ 708.146422] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 708.146603] env[61986]: INFO nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Took 0.03 seconds to destroy the instance on the hypervisor. [ 708.146932] env[61986]: DEBUG oslo.service.loopingcall [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.147089] env[61986]: DEBUG nova.compute.manager [-] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.147184] env[61986]: DEBUG nova.network.neutron [-] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.162243] env[61986]: DEBUG nova.network.neutron [-] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.223459] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.224645] env[61986]: ERROR nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Traceback (most recent call last): [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self.driver.spawn(context, instance, image_meta, [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] vm_ref = self.build_virtual_machine(instance, [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.224645] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] for vif in network_info: [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] return self._sync_wrapper(fn, *args, **kwargs) [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self.wait() [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self[:] = self._gt.wait() [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] return self._exit_event.wait() [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] current.throw(*self._exc) [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.225045] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] result = function(*args, **kwargs) [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] return func(*args, **kwargs) [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] raise e [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] nwinfo = self.network_api.allocate_for_instance( [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] created_port_ids = self._update_ports_for_instance( [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] with excutils.save_and_reraise_exception(): [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] self.force_reraise() [ 708.225477] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] raise self.value [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] updated_port = self._update_port( [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] _ensure_no_port_binding_failure(port) [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] raise exception.PortBindingFailed(port_id=port['id']) [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] nova.exception.PortBindingFailed: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. [ 708.225879] env[61986]: ERROR nova.compute.manager [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] [ 708.225879] env[61986]: DEBUG nova.compute.utils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.226815] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.991s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.227222] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.227222] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 708.227510] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.975s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.229309] env[61986]: INFO nova.compute.claims [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.232380] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Build of instance 4fa110e0-7235-4d62-a68b-c727b4a79908 was re-scheduled: Binding failed for port 36b058b4-6728-4495-84e4-6059c1ffed46, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.232865] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.233119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquiring lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.233273] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Acquired lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.233450] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 708.234963] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e72e1ba-d258-4504-8765-fa4817fbd506 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.244394] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f796628-860f-4ffb-94ee-26f15981b68e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.259960] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629cb84d-5fc3-4425-82b3-3867bd7bf413 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.266967] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f7eb49-c17f-4f34-bf24-2d0051c63429 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.297315] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181500MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 708.297486] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.416389] env[61986]: INFO nova.compute.manager [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] [instance: 35e41d19-8f1c-496d-9712-a1ac3c787c3b] Took 1.02 seconds to deallocate network for instance. [ 708.664545] env[61986]: DEBUG nova.network.neutron [-] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.764244] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.832928] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.167251] env[61986]: INFO nova.compute.manager [-] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Took 1.02 seconds to deallocate network for instance. [ 709.169792] env[61986]: DEBUG nova.compute.claims [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 709.170014] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.336529] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Releasing lock "refresh_cache-4fa110e0-7235-4d62-a68b-c727b4a79908" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.336836] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 709.337070] env[61986]: DEBUG nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.337258] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.351819] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.440601] env[61986]: INFO nova.scheduler.client.report [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Deleted allocations for instance 35e41d19-8f1c-496d-9712-a1ac3c787c3b [ 709.586978] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf2d7de-5482-4b05-8a83-e502f96b8604 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.597018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3764fb-67c3-4f68-aec7-d2e912f1e31d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.625694] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe2b982-45be-4749-8241-2664d9bde21b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.632947] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86da3717-993a-4c4f-aa11-0d7e18a376be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.645559] env[61986]: DEBUG nova.compute.provider_tree [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.854333] env[61986]: DEBUG nova.network.neutron [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.948408] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7ca57278-79a4-4c59-beec-8b98c3291861 tempest-InstanceActionsTestJSON-1664504763 tempest-InstanceActionsTestJSON-1664504763-project-member] Lock "35e41d19-8f1c-496d-9712-a1ac3c787c3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.158s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.148994] env[61986]: DEBUG nova.scheduler.client.report [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.357204] env[61986]: INFO nova.compute.manager [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] [instance: 4fa110e0-7235-4d62-a68b-c727b4a79908] Took 1.02 seconds to deallocate network for instance. [ 710.450854] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.656720] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.656720] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.658257] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.612s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.982850] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.163657] env[61986]: DEBUG nova.compute.utils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.168223] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.168428] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.228736] env[61986]: DEBUG nova.policy [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c6dfc9a8e294015812f4926e07593e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '040056883ae342589d380bc37eaec5a3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 711.386449] env[61986]: INFO nova.scheduler.client.report [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Deleted allocations for instance 4fa110e0-7235-4d62-a68b-c727b4a79908 [ 711.561018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89917e29-1efa-4b00-a3c4-c8b8a594afe4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.567858] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68961a14-12f2-456d-b592-06e7a449479d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.602857] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6fea8c-9cbb-4af4-89fc-43a9e3216f92 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.615019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c294543-29c2-4630-80b4-76c78460c685 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.626457] env[61986]: DEBUG nova.compute.provider_tree [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.665140] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Successfully created port: 3140ca67-9a97-45f8-95c5-1b6bf55bf177 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.668804] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.896327] env[61986]: DEBUG oslo_concurrency.lockutils [None req-292931fe-d987-48fc-92be-abc07b6db4d8 tempest-ServerExternalEventsTest-1201299396 tempest-ServerExternalEventsTest-1201299396-project-member] Lock "4fa110e0-7235-4d62-a68b-c727b4a79908" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.895s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.131301] env[61986]: DEBUG nova.scheduler.client.report [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.400016] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 712.505174] env[61986]: DEBUG nova.compute.manager [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Received event network-changed-3140ca67-9a97-45f8-95c5-1b6bf55bf177 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.505174] env[61986]: DEBUG nova.compute.manager [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Refreshing instance network info cache due to event network-changed-3140ca67-9a97-45f8-95c5-1b6bf55bf177. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 712.505174] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] Acquiring lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.505174] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] Acquired lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.505174] env[61986]: DEBUG nova.network.neutron [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Refreshing network info cache for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 712.635474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.636156] env[61986]: ERROR nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Traceback (most recent call last): [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self.driver.spawn(context, instance, image_meta, [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] vm_ref = self.build_virtual_machine(instance, [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.636156] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] for vif in network_info: [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] return self._sync_wrapper(fn, *args, **kwargs) [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self.wait() [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self[:] = self._gt.wait() [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] return self._exit_event.wait() [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] current.throw(*self._exc) [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.636518] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] result = function(*args, **kwargs) [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] return func(*args, **kwargs) [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] raise e [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] nwinfo = self.network_api.allocate_for_instance( [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] created_port_ids = self._update_ports_for_instance( [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] with excutils.save_and_reraise_exception(): [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] self.force_reraise() [ 712.636885] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] raise self.value [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] updated_port = self._update_port( [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] _ensure_no_port_binding_failure(port) [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] raise exception.PortBindingFailed(port_id=port['id']) [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] nova.exception.PortBindingFailed: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. [ 712.637285] env[61986]: ERROR nova.compute.manager [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] [ 712.637285] env[61986]: DEBUG nova.compute.utils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.638191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.495s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.645495] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Build of instance b632a5d1-62b1-44b1-a82a-b1e761e0c4b8 was re-scheduled: Binding failed for port f6d8e712-dc02-4bea-85d3-8166d0a31740, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.645921] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.647553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquiring lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.647708] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Acquired lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.647872] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.681048] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.713185] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.713525] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.713525] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.713695] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.716030] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.716030] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.716030] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.716030] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.716030] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.716298] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.716298] env[61986]: DEBUG nova.virt.hardware [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.716298] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52720946-9c05-4ece-83e7-58b4786cb36a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.725349] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd62eec9-f568-4234-8d4d-dbb45a83c964 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.845019] env[61986]: ERROR nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 712.845019] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.845019] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.845019] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.845019] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.845019] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.845019] env[61986]: ERROR nova.compute.manager raise self.value [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.845019] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.845019] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.845019] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.845478] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.845478] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.845478] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 712.845478] env[61986]: ERROR nova.compute.manager [ 712.845478] env[61986]: Traceback (most recent call last): [ 712.845478] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.845478] env[61986]: listener.cb(fileno) [ 712.845478] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.845478] env[61986]: result = function(*args, **kwargs) [ 712.845478] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.845478] env[61986]: return func(*args, **kwargs) [ 712.845478] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.845478] env[61986]: raise e [ 712.845478] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.845478] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 712.845478] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.845478] env[61986]: created_port_ids = self._update_ports_for_instance( [ 712.845478] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.845478] env[61986]: with excutils.save_and_reraise_exception(): [ 712.845478] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.845478] env[61986]: self.force_reraise() [ 712.845478] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.845478] env[61986]: raise self.value [ 712.845478] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.845478] env[61986]: updated_port = self._update_port( [ 712.845478] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.845478] env[61986]: _ensure_no_port_binding_failure(port) [ 712.845478] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.845478] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.846261] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 712.846261] env[61986]: Removing descriptor: 17 [ 712.846261] env[61986]: ERROR nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Traceback (most recent call last): [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] yield resources [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self.driver.spawn(context, instance, image_meta, [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.846261] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] vm_ref = self.build_virtual_machine(instance, [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] for vif in network_info: [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return self._sync_wrapper(fn, *args, **kwargs) [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self.wait() [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self[:] = self._gt.wait() [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return self._exit_event.wait() [ 712.846674] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] result = hub.switch() [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return self.greenlet.switch() [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] result = function(*args, **kwargs) [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return func(*args, **kwargs) [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] raise e [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] nwinfo = self.network_api.allocate_for_instance( [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.846975] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] created_port_ids = self._update_ports_for_instance( [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] with excutils.save_and_reraise_exception(): [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self.force_reraise() [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] raise self.value [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] updated_port = self._update_port( [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] _ensure_no_port_binding_failure(port) [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.847291] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] raise exception.PortBindingFailed(port_id=port['id']) [ 712.847582] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 712.847582] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] [ 712.847582] env[61986]: INFO nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Terminating instance [ 712.848476] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.924578] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.046019] env[61986]: DEBUG nova.network.neutron [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.186057] env[61986]: DEBUG nova.network.neutron [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.193972] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.371173] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.570625] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf056ab-5959-44d2-b6f2-49eef078c0dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.578711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5945259e-d957-49a8-923f-771099fa3361 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.615016] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cb1557-8204-49d2-afd8-8b063e839c11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.621639] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6813e3d1-89ef-4e5b-a371-2bedb84966e0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.636684] env[61986]: DEBUG nova.compute.provider_tree [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.688276] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bba6faf-2d84-4702-ae39-c1e5f21b6b03 req-d6bfe88e-1e49-4c69-9f4a-bcbfa8da2c3b service nova] Releasing lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.688658] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquired lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.688837] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.874441] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Releasing lock "refresh_cache-b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.875042] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.875306] env[61986]: DEBUG nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.875522] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.897283] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.139409] env[61986]: DEBUG nova.scheduler.client.report [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.211671] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.380169] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.399655] env[61986]: DEBUG nova.network.neutron [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.560391] env[61986]: DEBUG nova.compute.manager [req-94756aca-c5ff-45d3-86f8-a5be8f8d8c47 req-b91b7713-9358-4af2-bd3f-606e33aba47b service nova] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Received event network-vif-deleted-3140ca67-9a97-45f8-95c5-1b6bf55bf177 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.645948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.008s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.646686] env[61986]: ERROR nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Traceback (most recent call last): [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self.driver.spawn(context, instance, image_meta, [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] vm_ref = self.build_virtual_machine(instance, [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.646686] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] for vif in network_info: [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return self._sync_wrapper(fn, *args, **kwargs) [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self.wait() [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self[:] = self._gt.wait() [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return self._exit_event.wait() [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] result = hub.switch() [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.647015] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return self.greenlet.switch() [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] result = function(*args, **kwargs) [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] return func(*args, **kwargs) [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] raise e [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] nwinfo = self.network_api.allocate_for_instance( [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] created_port_ids = self._update_ports_for_instance( [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] with excutils.save_and_reraise_exception(): [ 714.647343] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] self.force_reraise() [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] raise self.value [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] updated_port = self._update_port( [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] _ensure_no_port_binding_failure(port) [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] raise exception.PortBindingFailed(port_id=port['id']) [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] nova.exception.PortBindingFailed: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. [ 714.647638] env[61986]: ERROR nova.compute.manager [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] [ 714.647900] env[61986]: DEBUG nova.compute.utils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.649128] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.642s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.652556] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Build of instance cbdf7e34-57f9-4456-8cc7-f09c15e0f965 was re-scheduled: Binding failed for port b1498699-69a3-44e4-bd22-edfd4f7f3633, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.652966] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.653212] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquiring lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.653358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Acquired lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.653514] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.881307] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Releasing lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.881743] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.881935] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 714.882255] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b0193ca-27f4-4ace-b803-5fe1a4e92ea3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.896381] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68047bc1-b4eb-4beb-b625-8fce37a08974 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.906608] env[61986]: INFO nova.compute.manager [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] [instance: b632a5d1-62b1-44b1-a82a-b1e761e0c4b8] Took 1.03 seconds to deallocate network for instance. [ 714.920783] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2 could not be found. [ 714.920979] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 714.924489] env[61986]: INFO nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 714.924489] env[61986]: DEBUG oslo.service.loopingcall [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.924489] env[61986]: DEBUG nova.compute.manager [-] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.924489] env[61986]: DEBUG nova.network.neutron [-] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 714.947786] env[61986]: DEBUG nova.network.neutron [-] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.184122] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.344628] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.452881] env[61986]: DEBUG nova.network.neutron [-] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.643462] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333684b5-c203-4307-998a-b93729a6d387 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.650194] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c510844d-b5d6-46a3-af89-0897534998a1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.680041] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d25fe39-2317-43f3-bd05-2b13050e83b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.687539] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b90313-3381-4f69-ab53-fec2ce18b675 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.701937] env[61986]: DEBUG nova.compute.provider_tree [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.849022] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Releasing lock "refresh_cache-cbdf7e34-57f9-4456-8cc7-f09c15e0f965" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.849155] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.849340] env[61986]: DEBUG nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.849506] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 715.870707] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.943225] env[61986]: INFO nova.scheduler.client.report [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Deleted allocations for instance b632a5d1-62b1-44b1-a82a-b1e761e0c4b8 [ 715.958551] env[61986]: INFO nova.compute.manager [-] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Took 1.04 seconds to deallocate network for instance. [ 715.961452] env[61986]: DEBUG nova.compute.claims [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 715.961608] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.992945] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.993215] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.206140] env[61986]: DEBUG nova.scheduler.client.report [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.374584] env[61986]: DEBUG nova.network.neutron [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.454795] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13bf06f0-c851-4f8d-b792-6a2d8fb20bdb tempest-ServerActionsTestJSON-731931803 tempest-ServerActionsTestJSON-731931803-project-member] Lock "b632a5d1-62b1-44b1-a82a-b1e761e0c4b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.938s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.712137] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.713259] env[61986]: ERROR nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Traceback (most recent call last): [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self.driver.spawn(context, instance, image_meta, [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] vm_ref = self.build_virtual_machine(instance, [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.713259] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] for vif in network_info: [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return self._sync_wrapper(fn, *args, **kwargs) [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self.wait() [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self[:] = self._gt.wait() [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return self._exit_event.wait() [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] result = hub.switch() [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.713697] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return self.greenlet.switch() [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] result = function(*args, **kwargs) [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] return func(*args, **kwargs) [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] raise e [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] nwinfo = self.network_api.allocate_for_instance( [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] created_port_ids = self._update_ports_for_instance( [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] with excutils.save_and_reraise_exception(): [ 716.714056] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] self.force_reraise() [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] raise self.value [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] updated_port = self._update_port( [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] _ensure_no_port_binding_failure(port) [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] raise exception.PortBindingFailed(port_id=port['id']) [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] nova.exception.PortBindingFailed: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. [ 716.714352] env[61986]: ERROR nova.compute.manager [instance: bf115dff-5440-41ba-9689-6337f63dedeb] [ 716.714611] env[61986]: DEBUG nova.compute.utils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.714967] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.014s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.717676] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Build of instance bf115dff-5440-41ba-9689-6337f63dedeb was re-scheduled: Binding failed for port 71e43a41-e010-4585-ad34-d2e52d8730b7, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.718149] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 716.718419] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquiring lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.718508] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Acquired lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.718666] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.876358] env[61986]: INFO nova.compute.manager [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] [instance: cbdf7e34-57f9-4456-8cc7-f09c15e0f965] Took 1.03 seconds to deallocate network for instance. [ 716.957120] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.259152] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.377672] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.481559] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.606715] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35790c0-6bd3-4fc5-b116-0e4f85044cb8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.614661] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5577a0c0-205b-4bb3-bfc4-e657d7a98b3e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.645859] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6856e9-336f-47bf-9bb0-09fc2213a860 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.653463] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab4bd81-3c40-4472-a38f-733e89e35867 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.667445] env[61986]: DEBUG nova.compute.provider_tree [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.884451] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Releasing lock "refresh_cache-bf115dff-5440-41ba-9689-6337f63dedeb" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.884451] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.884451] env[61986]: DEBUG nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.884451] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.932165] env[61986]: INFO nova.scheduler.client.report [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Deleted allocations for instance cbdf7e34-57f9-4456-8cc7-f09c15e0f965 [ 717.939562] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.174261] env[61986]: DEBUG nova.scheduler.client.report [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.444141] env[61986]: DEBUG nova.network.neutron [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.446802] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f734defb-b836-4a0e-bc65-e4c41814f531 tempest-ServersTestFqdnHostnames-1319514299 tempest-ServersTestFqdnHostnames-1319514299-project-member] Lock "cbdf7e34-57f9-4456-8cc7-f09c15e0f965" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.193s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.682176] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.682176] env[61986]: ERROR nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Traceback (most recent call last): [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self.driver.spawn(context, instance, image_meta, [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.682176] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] vm_ref = self.build_virtual_machine(instance, [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] for vif in network_info: [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] return self._sync_wrapper(fn, *args, **kwargs) [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self.wait() [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self[:] = self._gt.wait() [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] return self._exit_event.wait() [ 718.682422] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] current.throw(*self._exc) [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] result = function(*args, **kwargs) [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] return func(*args, **kwargs) [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] raise e [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] nwinfo = self.network_api.allocate_for_instance( [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] created_port_ids = self._update_ports_for_instance( [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 718.682717] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] with excutils.save_and_reraise_exception(): [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] self.force_reraise() [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] raise self.value [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] updated_port = self._update_port( [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] _ensure_no_port_binding_failure(port) [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] raise exception.PortBindingFailed(port_id=port['id']) [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] nova.exception.PortBindingFailed: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. [ 718.683031] env[61986]: ERROR nova.compute.manager [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] [ 718.683314] env[61986]: DEBUG nova.compute.utils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.687014] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.089s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.687014] env[61986]: INFO nova.compute.claims [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.688551] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Build of instance 9f8b9aca-420f-4314-a466-0d6f3d191377 was re-scheduled: Binding failed for port 0d64eca2-65c6-4fe6-8d9b-e2af0a22ed43, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 718.690221] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 718.690494] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquiring lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.691610] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Acquired lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.691837] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.948732] env[61986]: INFO nova.compute.manager [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] [instance: bf115dff-5440-41ba-9689-6337f63dedeb] Took 1.07 seconds to deallocate network for instance. [ 718.953043] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.218999] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.435893] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.481701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.941016] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Releasing lock "refresh_cache-9f8b9aca-420f-4314-a466-0d6f3d191377" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.941271] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.941454] env[61986]: DEBUG nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.941619] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 719.979152] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.981101] env[61986]: INFO nova.scheduler.client.report [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Deleted allocations for instance bf115dff-5440-41ba-9689-6337f63dedeb [ 720.118018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290b6ceb-d935-4e42-806a-0d68966f5b66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.126134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69eeae3d-ce61-4018-ae4f-15e280929601 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.159647] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d779d1-6d43-4bbe-8f61-7782f021491d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.168850] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e45ca30-50a0-454e-b32f-6f8e0194adaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.188997] env[61986]: DEBUG nova.compute.provider_tree [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.331363] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "86c9e45c-b3f1-4004-9acc-190b11a4a926" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.332059] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.496170] env[61986]: DEBUG nova.network.neutron [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.504119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0c3d28d6-f5eb-4e8a-acea-7e2b1abc76e2 tempest-ServerDiagnosticsNegativeTest-1231512518 tempest-ServerDiagnosticsNegativeTest-1231512518-project-member] Lock "bf115dff-5440-41ba-9689-6337f63dedeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.236s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.692066] env[61986]: DEBUG nova.scheduler.client.report [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.001261] env[61986]: INFO nova.compute.manager [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] [instance: 9f8b9aca-420f-4314-a466-0d6f3d191377] Took 1.06 seconds to deallocate network for instance. [ 721.011205] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.197260] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.197967] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.200976] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.691s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.202352] env[61986]: INFO nova.compute.claims [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.567166] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.712110] env[61986]: DEBUG nova.compute.utils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.712110] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.712110] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.769189] env[61986]: DEBUG nova.policy [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7fed7bfd27b4aefbd2274fffada643c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a342081088a9477cbcf2ef3a476ba291', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 722.059459] env[61986]: INFO nova.scheduler.client.report [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Deleted allocations for instance 9f8b9aca-420f-4314-a466-0d6f3d191377 [ 722.214915] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.428493] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Successfully created port: 5634e486-a2ab-4523-992a-fd0379b1f0db {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.571408] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7276bfb9-f7ac-47c8-baa5-4254eed5b657 tempest-AttachInterfacesV270Test-1445601591 tempest-AttachInterfacesV270Test-1445601591-project-member] Lock "9f8b9aca-420f-4314-a466-0d6f3d191377" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.196s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.626197] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8659e966-5c93-4e57-9353-3dad1c6faf55 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.634906] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb758b1-2a96-4d49-beaf-ad8c5af7930c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.672801] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a4a712-a27d-4279-9d6a-1dcd6ea6366b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.680354] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8a637e-e24e-44dc-9722-1769cd97a801 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.693508] env[61986]: DEBUG nova.compute.provider_tree [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.078783] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.197082] env[61986]: DEBUG nova.scheduler.client.report [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.224241] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.260833] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.261593] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.261839] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.262021] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.262181] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.262332] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.262534] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.262693] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.262933] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.263120] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.263295] env[61986]: DEBUG nova.virt.hardware [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.264513] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afec0fb1-b7da-4f71-ab00-f4cfc03e28c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.274176] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72dc8e2a-04d5-4729-be18-ee16a683f332 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.606277] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.704729] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.704729] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.706738] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.409s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.211660] env[61986]: DEBUG nova.compute.utils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.217403] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.217843] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 724.275351] env[61986]: DEBUG nova.policy [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7fed7bfd27b4aefbd2274fffada643c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a342081088a9477cbcf2ef3a476ba291', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 724.381246] env[61986]: DEBUG nova.compute.manager [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Received event network-changed-5634e486-a2ab-4523-992a-fd0379b1f0db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 724.381246] env[61986]: DEBUG nova.compute.manager [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Refreshing instance network info cache due to event network-changed-5634e486-a2ab-4523-992a-fd0379b1f0db. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 724.381246] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] Acquiring lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.381246] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] Acquired lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.381246] env[61986]: DEBUG nova.network.neutron [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Refreshing network info cache for port 5634e486-a2ab-4523-992a-fd0379b1f0db {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.656643] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Successfully created port: 1e949011-79f1-4aae-99bf-a73b87b96104 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.718745] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.767689] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a2c58c0a-785b-42a2-9960-0b3e4ff86d96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.769888] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.769888] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance fec98558-ff3a-4d65-a650-42b6c3ab9e18 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.769888] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 18f91633-5899-4f6d-addb-03c7ec39bc2c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 724.952898] env[61986]: DEBUG nova.network.neutron [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.083252] env[61986]: DEBUG nova.network.neutron [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.131994] env[61986]: ERROR nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 725.131994] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.131994] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.131994] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.131994] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.131994] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.131994] env[61986]: ERROR nova.compute.manager raise self.value [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.131994] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.131994] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.131994] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.134023] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.134023] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.134023] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 725.134023] env[61986]: ERROR nova.compute.manager [ 725.134023] env[61986]: Traceback (most recent call last): [ 725.134023] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.134023] env[61986]: listener.cb(fileno) [ 725.134023] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.134023] env[61986]: result = function(*args, **kwargs) [ 725.134023] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.134023] env[61986]: return func(*args, **kwargs) [ 725.134023] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.134023] env[61986]: raise e [ 725.134023] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.134023] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 725.134023] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.134023] env[61986]: created_port_ids = self._update_ports_for_instance( [ 725.134023] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.134023] env[61986]: with excutils.save_and_reraise_exception(): [ 725.134023] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.134023] env[61986]: self.force_reraise() [ 725.134023] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.134023] env[61986]: raise self.value [ 725.134023] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.134023] env[61986]: updated_port = self._update_port( [ 725.134023] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.134023] env[61986]: _ensure_no_port_binding_failure(port) [ 725.134023] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.134023] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.134757] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 725.134757] env[61986]: Removing descriptor: 14 [ 725.134757] env[61986]: ERROR nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Traceback (most recent call last): [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] yield resources [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self.driver.spawn(context, instance, image_meta, [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.134757] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] vm_ref = self.build_virtual_machine(instance, [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] for vif in network_info: [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return self._sync_wrapper(fn, *args, **kwargs) [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self.wait() [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self[:] = self._gt.wait() [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return self._exit_event.wait() [ 725.135090] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] result = hub.switch() [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return self.greenlet.switch() [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] result = function(*args, **kwargs) [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return func(*args, **kwargs) [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] raise e [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] nwinfo = self.network_api.allocate_for_instance( [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.135414] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] created_port_ids = self._update_ports_for_instance( [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] with excutils.save_and_reraise_exception(): [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self.force_reraise() [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] raise self.value [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] updated_port = self._update_port( [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] _ensure_no_port_binding_failure(port) [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.135771] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] raise exception.PortBindingFailed(port_id=port['id']) [ 725.136091] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 725.136091] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] [ 725.136091] env[61986]: INFO nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Terminating instance [ 725.140266] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.273058] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 213957b6-77a4-48fe-8bf5-f03bb67510c6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 725.591798] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ad4bf1f-38cf-41a4-a1f2-a045057ea12f req-f4de1a09-7ac3-41b1-9461-bfd3110094bb service nova] Releasing lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.592187] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquired lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.592378] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.743098] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.775560] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.775810] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.775961] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.776197] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.776353] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.776495] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.776699] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.776853] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.777022] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.777190] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.777398] env[61986]: DEBUG nova.virt.hardware [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.779223] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance dd239b21-fdb2-48fb-acf4-18242a98cd1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 725.781010] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b22dfd9-03bd-496b-bb5d-e6ef5faa3718 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.786099] env[61986]: DEBUG nova.compute.manager [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Received event network-changed-1e949011-79f1-4aae-99bf-a73b87b96104 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.786298] env[61986]: DEBUG nova.compute.manager [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Refreshing instance network info cache due to event network-changed-1e949011-79f1-4aae-99bf-a73b87b96104. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 725.786788] env[61986]: DEBUG oslo_concurrency.lockutils [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] Acquiring lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.786842] env[61986]: DEBUG oslo_concurrency.lockutils [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] Acquired lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.786999] env[61986]: DEBUG nova.network.neutron [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Refreshing network info cache for port 1e949011-79f1-4aae-99bf-a73b87b96104 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.795681] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b997288-d1cd-4c71-8863-c6608b4927df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.834379] env[61986]: ERROR nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 725.834379] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.834379] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.834379] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.834379] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.834379] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.834379] env[61986]: ERROR nova.compute.manager raise self.value [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.834379] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.834379] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.834379] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.834917] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.834917] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.834917] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 725.834917] env[61986]: ERROR nova.compute.manager [ 725.834917] env[61986]: Traceback (most recent call last): [ 725.834917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.834917] env[61986]: listener.cb(fileno) [ 725.834917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.834917] env[61986]: result = function(*args, **kwargs) [ 725.834917] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.834917] env[61986]: return func(*args, **kwargs) [ 725.834917] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.834917] env[61986]: raise e [ 725.834917] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.834917] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 725.834917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.834917] env[61986]: created_port_ids = self._update_ports_for_instance( [ 725.834917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.834917] env[61986]: with excutils.save_and_reraise_exception(): [ 725.834917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.834917] env[61986]: self.force_reraise() [ 725.834917] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.834917] env[61986]: raise self.value [ 725.834917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.834917] env[61986]: updated_port = self._update_port( [ 725.834917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.834917] env[61986]: _ensure_no_port_binding_failure(port) [ 725.834917] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.834917] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.835578] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 725.835578] env[61986]: Removing descriptor: 17 [ 725.835578] env[61986]: ERROR nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Traceback (most recent call last): [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] yield resources [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self.driver.spawn(context, instance, image_meta, [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.835578] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] vm_ref = self.build_virtual_machine(instance, [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] for vif in network_info: [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return self._sync_wrapper(fn, *args, **kwargs) [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self.wait() [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self[:] = self._gt.wait() [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return self._exit_event.wait() [ 725.835870] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] result = hub.switch() [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return self.greenlet.switch() [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] result = function(*args, **kwargs) [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return func(*args, **kwargs) [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] raise e [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] nwinfo = self.network_api.allocate_for_instance( [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.836207] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] created_port_ids = self._update_ports_for_instance( [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] with excutils.save_and_reraise_exception(): [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self.force_reraise() [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] raise self.value [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] updated_port = self._update_port( [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] _ensure_no_port_binding_failure(port) [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.836503] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] raise exception.PortBindingFailed(port_id=port['id']) [ 725.836772] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 725.836772] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] [ 725.836772] env[61986]: INFO nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Terminating instance [ 725.837891] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.110727] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.213796] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.285266] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 7e5fadf3-dd41-402c-9878-2d11295203ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 726.316661] env[61986]: DEBUG nova.network.neutron [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.413036] env[61986]: DEBUG nova.compute.manager [req-3633f908-9ad6-4bfa-84af-3ab925cc167b req-a4e0d579-810b-4dec-a2a8-be2ab4fc5abb service nova] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Received event network-vif-deleted-5634e486-a2ab-4523-992a-fd0379b1f0db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.449768] env[61986]: DEBUG nova.network.neutron [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.719395] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Releasing lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.719395] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.719395] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 726.719395] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46886ff1-10ff-451e-aa85-db732094c0ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.729440] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3024163f-4777-4292-9695-1e4d7eea7176 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.759574] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fec98558-ff3a-4d65-a650-42b6c3ab9e18 could not be found. [ 726.759953] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.760411] env[61986]: INFO nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.760730] env[61986]: DEBUG oslo.service.loopingcall [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.760963] env[61986]: DEBUG nova.compute.manager [-] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.761077] env[61986]: DEBUG nova.network.neutron [-] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.778639] env[61986]: DEBUG nova.network.neutron [-] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.792122] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 198f36d0-8be4-4885-9c0b-e85d6a0cbb22 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 726.952420] env[61986]: DEBUG oslo_concurrency.lockutils [req-08579ede-5df9-4041-a0c7-1b341e2a2fb4 req-840690f9-a936-43c3-8516-487d1d487b6a service nova] Releasing lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.952851] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquired lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.953039] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.284252] env[61986]: DEBUG nova.network.neutron [-] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.299470] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 77f4d9bd-c758-44c3-a17c-1efa8c395eb1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 727.488833] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.612530] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.657577] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.657806] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.787809] env[61986]: INFO nova.compute.manager [-] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Took 1.03 seconds to deallocate network for instance. [ 727.789255] env[61986]: DEBUG nova.compute.claims [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 727.789444] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.804198] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 55102fef-de7c-4505-a192-39568e08c096 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 727.808303] env[61986]: DEBUG nova.compute.manager [req-4b28b43f-2e0d-46f6-bd98-40081ebc88e1 req-9938a2f5-e13e-4b5b-88cc-98f3fb96e3fa service nova] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Received event network-vif-deleted-1e949011-79f1-4aae-99bf-a73b87b96104 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.116055] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Releasing lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.116055] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.116055] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 728.116055] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d547d255-afe2-4a32-b9d2-ae8434bae1e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.123103] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe2ce68-e48a-4b0f-8270-4ae003b3c80c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.150433] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 18f91633-5899-4f6d-addb-03c7ec39bc2c could not be found. [ 728.150669] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 728.150852] env[61986]: INFO nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 728.151247] env[61986]: DEBUG oslo.service.loopingcall [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.151611] env[61986]: DEBUG nova.compute.manager [-] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.151897] env[61986]: DEBUG nova.network.neutron [-] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 728.290790] env[61986]: DEBUG nova.network.neutron [-] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.307947] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 8bc12f2b-a632-4ffe-9d55-1a838e94931b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 728.782906] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.783155] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.793426] env[61986]: DEBUG nova.network.neutron [-] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.810575] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 2377b482-52e7-4054-8d5b-8f00587c3371 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 729.298049] env[61986]: INFO nova.compute.manager [-] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Took 1.14 seconds to deallocate network for instance. [ 729.299816] env[61986]: DEBUG nova.compute.claims [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 729.299816] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.314606] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa12919e-3891-4da9-a280-4155137864e1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 729.817162] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 730.320994] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6c62a1e9-d60c-4720-a554-951bea4b2e18 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 730.658922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.659170] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.824326] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 40de5f22-0de2-466a-91ab-dcb6ec586dad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.329032] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6e0697d4-d862-497c-9bdd-dd1ef2d4272b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 731.831263] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 977799b4-2793-4513-9447-483146fc7ac4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.334067] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 4bb588b3-c5e1-4619-b77b-6734d6516170 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 732.837269] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance c1a6ffac-99bd-492c-99e5-f7c46b352d8e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 733.341178] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ce09db77-1eac-4228-a4f8-228a3b7e69d9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 733.845044] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 8438d2a7-975c-4d1f-8906-f960bfadfe39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 734.348468] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance eaa61f81-2f9b-4d1c-bab8-4363fb71e936 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 734.852305] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6f20dd8e-9289-458d-bed6-f2ef9daaa917 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.355587] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ebf5fee5-7f1b-4537-aec3-77a8a963670b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.858596] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 86c9e45c-b3f1-4004-9acc-190b11a4a926 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 735.858909] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 735.859113] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 736.202419] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a2d956-8510-4289-b75c-9e3843180e86 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.209714] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93fb13ed-d727-46e3-a8be-3eda9ec918ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.238859] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d8219f-6e55-4737-9608-33b1e184b9c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.246027] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5a4281-80a1-4709-8b97-30f1eb0a8481 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.259178] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.762050] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.267722] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 737.267973] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.561s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.268283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.098s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.322919] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb28b5da-f758-44ff-bade-122d37e5bdcf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.332647] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c530ca56-5c06-4395-af1f-3c97fd708c8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.369971] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e43f65-8fb9-4629-ae00-245f40fd6224 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.380477] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cd1632-a01b-428c-8979-6c44af0e42f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.398028] env[61986]: DEBUG nova.compute.provider_tree [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.900294] env[61986]: DEBUG nova.scheduler.client.report [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.405732] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.137s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.406448] env[61986]: ERROR nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Traceback (most recent call last): [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self.driver.spawn(context, instance, image_meta, [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] vm_ref = self.build_virtual_machine(instance, [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.406448] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] for vif in network_info: [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] return self._sync_wrapper(fn, *args, **kwargs) [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self.wait() [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self[:] = self._gt.wait() [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] return self._exit_event.wait() [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] current.throw(*self._exc) [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.406751] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] result = function(*args, **kwargs) [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] return func(*args, **kwargs) [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] raise e [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] nwinfo = self.network_api.allocate_for_instance( [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] created_port_ids = self._update_ports_for_instance( [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] with excutils.save_and_reraise_exception(): [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] self.force_reraise() [ 739.407064] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] raise self.value [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] updated_port = self._update_port( [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] _ensure_no_port_binding_failure(port) [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] raise exception.PortBindingFailed(port_id=port['id']) [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] nova.exception.PortBindingFailed: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. [ 739.407483] env[61986]: ERROR nova.compute.manager [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] [ 739.407483] env[61986]: DEBUG nova.compute.utils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.409887] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.426s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.411325] env[61986]: INFO nova.compute.claims [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.414910] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Build of instance a2c58c0a-785b-42a2-9960-0b3e4ff86d96 was re-scheduled: Binding failed for port 06a228ab-e02b-4472-a877-22ee30cb8d8e, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.415354] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.415580] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquiring lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.415722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Acquired lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.415878] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.935500] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.011071] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.514393] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Releasing lock "refresh_cache-a2c58c0a-785b-42a2-9960-0b3e4ff86d96" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.514620] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.514690] env[61986]: DEBUG nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.514852] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 740.532101] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.739290] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2c7c4b-dcb7-47ab-a0e1-119e22fb4a8a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.746622] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbad11f5-39fb-4df7-87fe-f61c4f68b7c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.775518] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba340464-3cc0-4655-804c-1f0178dd25a6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.782239] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b198eff2-e07f-4a49-9700-3225dae89c26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.795936] env[61986]: DEBUG nova.compute.provider_tree [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.034178] env[61986]: DEBUG nova.network.neutron [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.299530] env[61986]: DEBUG nova.scheduler.client.report [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.536745] env[61986]: INFO nova.compute.manager [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] [instance: a2c58c0a-785b-42a2-9960-0b3e4ff86d96] Took 1.02 seconds to deallocate network for instance. [ 741.804582] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.805124] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.807654] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.884s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.809829] env[61986]: INFO nova.compute.claims [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.315567] env[61986]: DEBUG nova.compute.utils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.319019] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.319146] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 742.357729] env[61986]: DEBUG nova.policy [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7fed7bfd27b4aefbd2274fffada643c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a342081088a9477cbcf2ef3a476ba291', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 742.578189] env[61986]: INFO nova.scheduler.client.report [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Deleted allocations for instance a2c58c0a-785b-42a2-9960-0b3e4ff86d96 [ 742.616802] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Successfully created port: 8d610236-808e-469e-b304-e781a4cc5fa2 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.822332] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.088901] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979d4171-575d-4b03-b051-57c38be8f4b7 tempest-ServersV294TestFqdnHostnames-2007494521 tempest-ServersV294TestFqdnHostnames-2007494521-project-member] Lock "a2c58c0a-785b-42a2-9960-0b3e4ff86d96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.676s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.195927] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987eeb03-1a6b-47a0-ac39-855fe1a4cdaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.206155] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335fdd30-c6de-4442-a64e-2db79732bf68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.239733] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6078d5-c1f2-46e5-b778-2acc24db4e72 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.247808] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc021fe-28e7-484a-9876-da28934361e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.264905] env[61986]: DEBUG nova.compute.provider_tree [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.527600] env[61986]: DEBUG nova.compute.manager [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Received event network-changed-8d610236-808e-469e-b304-e781a4cc5fa2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 743.527801] env[61986]: DEBUG nova.compute.manager [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Refreshing instance network info cache due to event network-changed-8d610236-808e-469e-b304-e781a4cc5fa2. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 743.527999] env[61986]: DEBUG oslo_concurrency.lockutils [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] Acquiring lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.528153] env[61986]: DEBUG oslo_concurrency.lockutils [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] Acquired lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.528375] env[61986]: DEBUG nova.network.neutron [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Refreshing network info cache for port 8d610236-808e-469e-b304-e781a4cc5fa2 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.557833] env[61986]: ERROR nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 743.557833] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.557833] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.557833] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.557833] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.557833] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.557833] env[61986]: ERROR nova.compute.manager raise self.value [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.557833] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.557833] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.557833] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.558412] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.558412] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.558412] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 743.558412] env[61986]: ERROR nova.compute.manager [ 743.558412] env[61986]: Traceback (most recent call last): [ 743.558412] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.558412] env[61986]: listener.cb(fileno) [ 743.558412] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.558412] env[61986]: result = function(*args, **kwargs) [ 743.558412] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.558412] env[61986]: return func(*args, **kwargs) [ 743.558412] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.558412] env[61986]: raise e [ 743.558412] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.558412] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 743.558412] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.558412] env[61986]: created_port_ids = self._update_ports_for_instance( [ 743.558412] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.558412] env[61986]: with excutils.save_and_reraise_exception(): [ 743.558412] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.558412] env[61986]: self.force_reraise() [ 743.558412] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.558412] env[61986]: raise self.value [ 743.558412] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.558412] env[61986]: updated_port = self._update_port( [ 743.558412] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.558412] env[61986]: _ensure_no_port_binding_failure(port) [ 743.558412] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.558412] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.559098] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 743.559098] env[61986]: Removing descriptor: 17 [ 743.593350] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.767626] env[61986]: DEBUG nova.scheduler.client.report [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.830972] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.858413] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.858664] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.858820] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.858986] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.862869] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.863061] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.863288] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.863450] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.863617] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.863779] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.863950] env[61986]: DEBUG nova.virt.hardware [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.868025] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b0daf9-8a9c-4962-b26c-163b8f0cc2a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.873403] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ab0efe-1a97-4e15-ae0e-126b86235430 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.888287] env[61986]: ERROR nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Traceback (most recent call last): [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] yield resources [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self.driver.spawn(context, instance, image_meta, [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] vm_ref = self.build_virtual_machine(instance, [ 743.888287] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] for vif in network_info: [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] return self._sync_wrapper(fn, *args, **kwargs) [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self.wait() [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self[:] = self._gt.wait() [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] return self._exit_event.wait() [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.888702] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] current.throw(*self._exc) [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] result = function(*args, **kwargs) [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] return func(*args, **kwargs) [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] raise e [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] nwinfo = self.network_api.allocate_for_instance( [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] created_port_ids = self._update_ports_for_instance( [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] with excutils.save_and_reraise_exception(): [ 743.889018] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self.force_reraise() [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] raise self.value [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] updated_port = self._update_port( [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] _ensure_no_port_binding_failure(port) [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] raise exception.PortBindingFailed(port_id=port['id']) [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 743.889333] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] [ 743.889333] env[61986]: INFO nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Terminating instance [ 743.890927] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.050021] env[61986]: DEBUG nova.network.neutron [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.115130] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.123308] env[61986]: DEBUG nova.network.neutron [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.272150] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.272665] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.275373] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.314s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.627324] env[61986]: DEBUG oslo_concurrency.lockutils [req-57591869-4233-49f5-ab95-7765f28509e4 req-967b420d-3e85-4fda-940b-84af5ba5983d service nova] Releasing lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.627813] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquired lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.627999] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 744.780099] env[61986]: DEBUG nova.compute.utils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.788021] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.788021] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 744.835543] env[61986]: DEBUG nova.policy [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0175ce4758c64a7c88054f26ec42dfe9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '575f638c6e5041b6aa43d6c49ea88042', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 745.150428] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.166918] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d732e5-90eb-41df-b485-31e97e935e42 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.170485] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Successfully created port: b307ac55-fe5e-4042-84a3-f24a463c1434 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.176769] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619c7e85-d7a9-4ee4-a70a-efc11965ebfc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.218527] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84f12fd-cd7e-41ff-8d7b-7cc2e0bf086f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.225716] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1338d8c6-5501-4405-a8b2-4ae387058da9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.239442] env[61986]: DEBUG nova.compute.provider_tree [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.288121] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.302632] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.554339] env[61986]: DEBUG nova.compute.manager [req-c90c4d4e-5dde-42dc-af08-bce8adf28a59 req-aecec655-3786-4b7c-9280-111670d20e09 service nova] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Received event network-vif-deleted-8d610236-808e-469e-b304-e781a4cc5fa2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 745.744016] env[61986]: DEBUG nova.scheduler.client.report [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.808251] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Releasing lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.808251] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.808251] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 745.808251] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-234de60a-ba9b-4393-bb99-f83bcbe563ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.817763] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c09784c-4d84-4544-8624-384156b9ec17 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.844377] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 213957b6-77a4-48fe-8bf5-f03bb67510c6 could not be found. [ 745.844624] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 745.844802] env[61986]: INFO nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 745.845057] env[61986]: DEBUG oslo.service.loopingcall [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.845280] env[61986]: DEBUG nova.compute.manager [-] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.845387] env[61986]: DEBUG nova.network.neutron [-] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 745.867231] env[61986]: DEBUG nova.network.neutron [-] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.250203] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.250856] env[61986]: ERROR nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Traceback (most recent call last): [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self.driver.spawn(context, instance, image_meta, [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] vm_ref = self.build_virtual_machine(instance, [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.250856] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] for vif in network_info: [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return self._sync_wrapper(fn, *args, **kwargs) [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self.wait() [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self[:] = self._gt.wait() [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return self._exit_event.wait() [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] result = hub.switch() [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.251197] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return self.greenlet.switch() [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] result = function(*args, **kwargs) [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] return func(*args, **kwargs) [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] raise e [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] nwinfo = self.network_api.allocate_for_instance( [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] created_port_ids = self._update_ports_for_instance( [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] with excutils.save_and_reraise_exception(): [ 746.251560] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] self.force_reraise() [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] raise self.value [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] updated_port = self._update_port( [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] _ensure_no_port_binding_failure(port) [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] raise exception.PortBindingFailed(port_id=port['id']) [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] nova.exception.PortBindingFailed: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. [ 746.251918] env[61986]: ERROR nova.compute.manager [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] [ 746.252229] env[61986]: DEBUG nova.compute.utils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 746.252814] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.771s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.257128] env[61986]: INFO nova.compute.claims [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.260233] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Build of instance 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2 was re-scheduled: Binding failed for port 3140ca67-9a97-45f8-95c5-1b6bf55bf177, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 746.260518] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 746.260747] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.260891] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquired lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.261056] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 746.276086] env[61986]: ERROR nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 746.276086] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.276086] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.276086] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.276086] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.276086] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.276086] env[61986]: ERROR nova.compute.manager raise self.value [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.276086] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.276086] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.276086] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.276566] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.276566] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.276566] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 746.276566] env[61986]: ERROR nova.compute.manager [ 746.276566] env[61986]: Traceback (most recent call last): [ 746.276566] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.276566] env[61986]: listener.cb(fileno) [ 746.276566] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.276566] env[61986]: result = function(*args, **kwargs) [ 746.276566] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.276566] env[61986]: return func(*args, **kwargs) [ 746.276566] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.276566] env[61986]: raise e [ 746.276566] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.276566] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 746.276566] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.276566] env[61986]: created_port_ids = self._update_ports_for_instance( [ 746.276566] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.276566] env[61986]: with excutils.save_and_reraise_exception(): [ 746.276566] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.276566] env[61986]: self.force_reraise() [ 746.276566] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.276566] env[61986]: raise self.value [ 746.276566] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.276566] env[61986]: updated_port = self._update_port( [ 746.276566] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.276566] env[61986]: _ensure_no_port_binding_failure(port) [ 746.276566] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.276566] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.277256] env[61986]: nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 746.277256] env[61986]: Removing descriptor: 17 [ 746.304073] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.331107] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.331429] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.331587] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.331842] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.332044] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.332215] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.334568] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.334568] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.334568] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.334568] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.334568] env[61986]: DEBUG nova.virt.hardware [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.334866] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b849e0-8c07-485f-a88d-b2bdc3cfc4a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.345771] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9f691c-cba4-4d63-a555-a225092f4074 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.360061] env[61986]: ERROR nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Traceback (most recent call last): [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] yield resources [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self.driver.spawn(context, instance, image_meta, [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] vm_ref = self.build_virtual_machine(instance, [ 746.360061] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] for vif in network_info: [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] return self._sync_wrapper(fn, *args, **kwargs) [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self.wait() [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self[:] = self._gt.wait() [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] return self._exit_event.wait() [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.360379] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] current.throw(*self._exc) [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] result = function(*args, **kwargs) [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] return func(*args, **kwargs) [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] raise e [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] nwinfo = self.network_api.allocate_for_instance( [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] created_port_ids = self._update_ports_for_instance( [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] with excutils.save_and_reraise_exception(): [ 746.360691] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self.force_reraise() [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] raise self.value [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] updated_port = self._update_port( [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] _ensure_no_port_binding_failure(port) [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] raise exception.PortBindingFailed(port_id=port['id']) [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 746.361044] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] [ 746.361044] env[61986]: INFO nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Terminating instance [ 746.363763] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquiring lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.363763] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquired lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.363763] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 746.371018] env[61986]: DEBUG nova.network.neutron [-] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.784041] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.859370] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.872579] env[61986]: INFO nova.compute.manager [-] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Took 1.03 seconds to deallocate network for instance. [ 746.874556] env[61986]: DEBUG nova.compute.claims [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 746.874729] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.878478] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.949658] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.361494] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Releasing lock "refresh_cache-8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.361735] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 747.361917] env[61986]: DEBUG nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.362097] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 747.378564] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.451928] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Releasing lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.452381] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.452574] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 747.453737] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ea7b08f-ceb2-40a1-87e1-862df7ee54f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.463088] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209f9e3c-4038-4734-8d93-6e187d8d9cea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.488584] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dd239b21-fdb2-48fb-acf4-18242a98cd1f could not be found. [ 747.488882] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 747.489117] env[61986]: INFO nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 747.489421] env[61986]: DEBUG oslo.service.loopingcall [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.491979] env[61986]: DEBUG nova.compute.manager [-] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.492131] env[61986]: DEBUG nova.network.neutron [-] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 747.511190] env[61986]: DEBUG nova.network.neutron [-] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.588169] env[61986]: DEBUG nova.compute.manager [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Received event network-changed-b307ac55-fe5e-4042-84a3-f24a463c1434 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.588378] env[61986]: DEBUG nova.compute.manager [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Refreshing instance network info cache due to event network-changed-b307ac55-fe5e-4042-84a3-f24a463c1434. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 747.588588] env[61986]: DEBUG oslo_concurrency.lockutils [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] Acquiring lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.588882] env[61986]: DEBUG oslo_concurrency.lockutils [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] Acquired lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.588882] env[61986]: DEBUG nova.network.neutron [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Refreshing network info cache for port b307ac55-fe5e-4042-84a3-f24a463c1434 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 747.648647] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9850fe25-8979-4ca0-9155-5e2cfb892a2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.656270] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa0da95-eb66-4aef-8d7d-8a683f5382e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.690322] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686e75ca-8224-4d02-ab5a-10903a42d1b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.700304] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c162cd-63e6-4310-89f0-4964692f7799 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.714543] env[61986]: DEBUG nova.compute.provider_tree [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.881385] env[61986]: DEBUG nova.network.neutron [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.017368] env[61986]: DEBUG nova.network.neutron [-] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.116427] env[61986]: DEBUG nova.network.neutron [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.216910] env[61986]: DEBUG nova.scheduler.client.report [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.241404] env[61986]: DEBUG nova.network.neutron [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.384552] env[61986]: INFO nova.compute.manager [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2] Took 1.02 seconds to deallocate network for instance. [ 748.521864] env[61986]: INFO nova.compute.manager [-] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Took 1.03 seconds to deallocate network for instance. [ 748.522915] env[61986]: DEBUG nova.compute.claims [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 748.523334] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.721878] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.722627] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.725024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.243s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.726907] env[61986]: INFO nova.compute.claims [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.744651] env[61986]: DEBUG oslo_concurrency.lockutils [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] Releasing lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.744918] env[61986]: DEBUG nova.compute.manager [req-f83cead4-bc18-4a1d-84b4-f22fc8880924 req-4ddcd0db-e2d3-4cc3-9bd7-54f4ab36f5d8 service nova] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Received event network-vif-deleted-b307ac55-fe5e-4042-84a3-f24a463c1434 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.238560] env[61986]: DEBUG nova.compute.utils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.240159] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.240296] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 749.289500] env[61986]: DEBUG nova.policy [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9bfa1d070ad49edbfd1390a8a2d3bb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55d5731446c44ba9a45cd1d25c4a3eb8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 749.430754] env[61986]: INFO nova.scheduler.client.report [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Deleted allocations for instance 8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2 [ 749.623383] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Successfully created port: 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.744741] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.946387] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5ebeea63-349b-4437-b6ed-5d3c1d20ca86 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "8adfe7b8-0e4d-45ca-9f64-b2ed33e85ec2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.822s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.155243] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4dea5eb-6072-493c-ba7c-13042dfa687b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.163834] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618a8077-0756-492b-9c7f-509cba9144cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.195943] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5503da5-8d3c-4fd1-b402-84e38cd26ce8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.203720] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb5dac2-3da5-4ee5-9f72-3453d09a2a6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.217835] env[61986]: DEBUG nova.compute.provider_tree [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.420712] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "182479b8-f72f-4395-99a4-af0d6f91f7d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.420712] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.451880] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.661143] env[61986]: DEBUG nova.compute.manager [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Received event network-changed-5b75aa7a-be06-41e3-a4a3-42e9e5b541d3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 750.661143] env[61986]: DEBUG nova.compute.manager [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Refreshing instance network info cache due to event network-changed-5b75aa7a-be06-41e3-a4a3-42e9e5b541d3. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 750.661143] env[61986]: DEBUG oslo_concurrency.lockutils [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] Acquiring lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.661374] env[61986]: DEBUG oslo_concurrency.lockutils [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] Acquired lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.661374] env[61986]: DEBUG nova.network.neutron [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Refreshing network info cache for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 750.693143] env[61986]: ERROR nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 750.693143] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.693143] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.693143] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.693143] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.693143] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.693143] env[61986]: ERROR nova.compute.manager raise self.value [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.693143] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.693143] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.693143] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.693626] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.693626] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.693626] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 750.693626] env[61986]: ERROR nova.compute.manager [ 750.693626] env[61986]: Traceback (most recent call last): [ 750.693626] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.693626] env[61986]: listener.cb(fileno) [ 750.693626] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.693626] env[61986]: result = function(*args, **kwargs) [ 750.693626] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.693626] env[61986]: return func(*args, **kwargs) [ 750.693626] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.693626] env[61986]: raise e [ 750.693626] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.693626] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 750.693626] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.693626] env[61986]: created_port_ids = self._update_ports_for_instance( [ 750.693626] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.693626] env[61986]: with excutils.save_and_reraise_exception(): [ 750.693626] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.693626] env[61986]: self.force_reraise() [ 750.693626] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.693626] env[61986]: raise self.value [ 750.693626] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.693626] env[61986]: updated_port = self._update_port( [ 750.693626] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.693626] env[61986]: _ensure_no_port_binding_failure(port) [ 750.693626] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.693626] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.694439] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 750.694439] env[61986]: Removing descriptor: 17 [ 750.720783] env[61986]: DEBUG nova.scheduler.client.report [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.755419] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.786767] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.786919] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.787046] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.787226] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.787372] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.787517] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.787713] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.787868] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.788038] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.788203] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.788370] env[61986]: DEBUG nova.virt.hardware [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.789306] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd24de52-d73a-4f12-9614-a263ec9a65c8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.797711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aeb720e-e2db-4fdb-8dd4-2519de43f73c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.812684] env[61986]: ERROR nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Traceback (most recent call last): [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] yield resources [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self.driver.spawn(context, instance, image_meta, [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] vm_ref = self.build_virtual_machine(instance, [ 750.812684] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] for vif in network_info: [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] return self._sync_wrapper(fn, *args, **kwargs) [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self.wait() [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self[:] = self._gt.wait() [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] return self._exit_event.wait() [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.813061] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] current.throw(*self._exc) [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] result = function(*args, **kwargs) [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] return func(*args, **kwargs) [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] raise e [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] nwinfo = self.network_api.allocate_for_instance( [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] created_port_ids = self._update_ports_for_instance( [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] with excutils.save_and_reraise_exception(): [ 750.813421] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self.force_reraise() [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] raise self.value [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] updated_port = self._update_port( [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] _ensure_no_port_binding_failure(port) [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] raise exception.PortBindingFailed(port_id=port['id']) [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 750.813742] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] [ 750.813742] env[61986]: INFO nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Terminating instance [ 750.815020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquiring lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.970615] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.183372] env[61986]: DEBUG nova.network.neutron [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.194836] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.195159] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.230725] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.231059] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.233606] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.667s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.235706] env[61986]: INFO nova.compute.claims [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.279367] env[61986]: DEBUG nova.network.neutron [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.741577] env[61986]: DEBUG nova.compute.utils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.743346] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.743513] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 751.781986] env[61986]: DEBUG oslo_concurrency.lockutils [req-1e818aaa-041a-475e-b088-916eb0770eb0 req-d39333bc-aac4-493a-901e-27a664cbb91a service nova] Releasing lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.783676] env[61986]: DEBUG nova.policy [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a79defe9334d45e396e1bc6f68da7da2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30cffc57b42847d8a13ac2347ea2ade7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 751.785217] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquired lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.785391] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.069745] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Successfully created port: 925120df-b890-4643-91ee-74efd811ab1c {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.247679] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.323396] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.438581] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.626936] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a38c117-5848-4a94-b199-44b8348f0be1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.635111] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a70e70e-4a9f-45e5-99b2-9964f5413a45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.664940] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eb35f1-15fe-4c13-a6d8-b288e6b92f08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.672447] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f9b326-107e-4efc-bfd4-289a2ab471e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.685206] env[61986]: DEBUG nova.compute.provider_tree [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.691331] env[61986]: DEBUG nova.compute.manager [req-8a3852a1-d93a-470e-8027-f7ee4e97a560 req-f8a076f6-59f8-4011-8939-bac625b945cf service nova] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Received event network-vif-deleted-5b75aa7a-be06-41e3-a4a3-42e9e5b541d3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.946993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Releasing lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.946993] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.946993] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.946993] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a97dd00-ff5d-492b-be9f-643661f20068 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.948108] env[61986]: ERROR nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 752.948108] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.948108] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.948108] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.948108] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.948108] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.948108] env[61986]: ERROR nova.compute.manager raise self.value [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.948108] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 752.948108] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.948108] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 752.949258] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.949258] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 752.949258] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 752.949258] env[61986]: ERROR nova.compute.manager [ 752.949258] env[61986]: Traceback (most recent call last): [ 752.949258] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 752.949258] env[61986]: listener.cb(fileno) [ 752.949258] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.949258] env[61986]: result = function(*args, **kwargs) [ 752.949258] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 752.949258] env[61986]: return func(*args, **kwargs) [ 752.949258] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.949258] env[61986]: raise e [ 752.949258] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.949258] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 752.949258] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.949258] env[61986]: created_port_ids = self._update_ports_for_instance( [ 752.949258] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.949258] env[61986]: with excutils.save_and_reraise_exception(): [ 752.949258] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.949258] env[61986]: self.force_reraise() [ 752.949258] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.949258] env[61986]: raise self.value [ 752.949258] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.949258] env[61986]: updated_port = self._update_port( [ 752.949258] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.949258] env[61986]: _ensure_no_port_binding_failure(port) [ 752.949258] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.949258] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 752.950152] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 752.950152] env[61986]: Removing descriptor: 17 [ 752.954980] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b016d5a-c5e1-4593-972b-2ed334765684 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.975385] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7e5fadf3-dd41-402c-9878-2d11295203ac could not be found. [ 752.975595] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.975769] env[61986]: INFO nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.976040] env[61986]: DEBUG oslo.service.loopingcall [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.976238] env[61986]: DEBUG nova.compute.manager [-] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.976331] env[61986]: DEBUG nova.network.neutron [-] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.990874] env[61986]: DEBUG nova.network.neutron [-] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.188744] env[61986]: DEBUG nova.scheduler.client.report [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.260250] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.283189] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.283442] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.283594] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.283768] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.283905] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.284055] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.284275] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.284407] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.284564] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.284818] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.284992] env[61986]: DEBUG nova.virt.hardware [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.285852] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2560d476-8bf6-4ec0-9b27-8f143e484d8c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.293714] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9e368b-382d-47d6-889f-0183800f60d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.307040] env[61986]: ERROR nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Traceback (most recent call last): [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] yield resources [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self.driver.spawn(context, instance, image_meta, [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] vm_ref = self.build_virtual_machine(instance, [ 753.307040] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] for vif in network_info: [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] return self._sync_wrapper(fn, *args, **kwargs) [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self.wait() [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self[:] = self._gt.wait() [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] return self._exit_event.wait() [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.307457] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] current.throw(*self._exc) [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] result = function(*args, **kwargs) [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] return func(*args, **kwargs) [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] raise e [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] nwinfo = self.network_api.allocate_for_instance( [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] created_port_ids = self._update_ports_for_instance( [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] with excutils.save_and_reraise_exception(): [ 753.307900] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self.force_reraise() [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] raise self.value [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] updated_port = self._update_port( [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] _ensure_no_port_binding_failure(port) [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] raise exception.PortBindingFailed(port_id=port['id']) [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 753.308293] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] [ 753.308293] env[61986]: INFO nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Terminating instance [ 753.309553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.309770] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquired lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.309915] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.494097] env[61986]: DEBUG nova.network.neutron [-] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.693886] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.694588] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 753.697266] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.091s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.698727] env[61986]: INFO nova.compute.claims [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.827736] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.909712] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.996596] env[61986]: INFO nova.compute.manager [-] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Took 1.02 seconds to deallocate network for instance. [ 753.999134] env[61986]: DEBUG nova.compute.claims [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 753.999314] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.203246] env[61986]: DEBUG nova.compute.utils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.208395] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 754.208395] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 754.267357] env[61986]: DEBUG nova.policy [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '415b0e7dfa1541bfba60c687002c440c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5163e73250e14ba89e003d97bdf39ba2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 754.412760] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Releasing lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.413216] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 754.413408] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 754.413744] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d754ef8-386f-4fe2-8890-c30cbe6ddc93 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.422713] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3beb4150-94de-4e25-92d0-66b3e18f6dc8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.444733] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 198f36d0-8be4-4885-9c0b-e85d6a0cbb22 could not be found. [ 754.445057] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.445281] env[61986]: INFO nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Took 0.03 seconds to destroy the instance on the hypervisor. [ 754.445606] env[61986]: DEBUG oslo.service.loopingcall [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.445969] env[61986]: DEBUG nova.compute.manager [-] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.446043] env[61986]: DEBUG nova.network.neutron [-] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.476486] env[61986]: DEBUG nova.network.neutron [-] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.529066] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Successfully created port: b813b6da-fc0a-4458-b1fb-9e8992b72c92 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.708285] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 754.716485] env[61986]: DEBUG nova.compute.manager [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Received event network-changed-925120df-b890-4643-91ee-74efd811ab1c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 754.716796] env[61986]: DEBUG nova.compute.manager [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Refreshing instance network info cache due to event network-changed-925120df-b890-4643-91ee-74efd811ab1c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 754.717015] env[61986]: DEBUG oslo_concurrency.lockutils [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] Acquiring lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.717170] env[61986]: DEBUG oslo_concurrency.lockutils [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] Acquired lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.717332] env[61986]: DEBUG nova.network.neutron [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Refreshing network info cache for port 925120df-b890-4643-91ee-74efd811ab1c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 754.982606] env[61986]: DEBUG nova.network.neutron [-] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.137211] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287df4e2-5cde-4d90-bd5a-8d911fc6b112 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.145036] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1480f20-b229-4568-bf0e-9789762bf004 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.176114] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbafb8e-a882-4733-8a44-7bf7ecc7df3c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.183940] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c299ca-b15e-483f-ad9e-b15418102439 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.200126] env[61986]: DEBUG nova.compute.provider_tree [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.240463] env[61986]: DEBUG nova.network.neutron [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.324974] env[61986]: DEBUG nova.network.neutron [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.438545] env[61986]: ERROR nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 755.438545] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.438545] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.438545] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.438545] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.438545] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.438545] env[61986]: ERROR nova.compute.manager raise self.value [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.438545] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 755.438545] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.438545] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 755.439045] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.439045] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 755.439045] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 755.439045] env[61986]: ERROR nova.compute.manager [ 755.439045] env[61986]: Traceback (most recent call last): [ 755.439045] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 755.439045] env[61986]: listener.cb(fileno) [ 755.439045] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.439045] env[61986]: result = function(*args, **kwargs) [ 755.439045] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.439045] env[61986]: return func(*args, **kwargs) [ 755.439045] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.439045] env[61986]: raise e [ 755.439045] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.439045] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 755.439045] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.439045] env[61986]: created_port_ids = self._update_ports_for_instance( [ 755.439045] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.439045] env[61986]: with excutils.save_and_reraise_exception(): [ 755.439045] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.439045] env[61986]: self.force_reraise() [ 755.439045] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.439045] env[61986]: raise self.value [ 755.439045] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.439045] env[61986]: updated_port = self._update_port( [ 755.439045] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.439045] env[61986]: _ensure_no_port_binding_failure(port) [ 755.439045] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.439045] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 755.439922] env[61986]: nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 755.439922] env[61986]: Removing descriptor: 17 [ 755.485267] env[61986]: INFO nova.compute.manager [-] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Took 1.04 seconds to deallocate network for instance. [ 755.487896] env[61986]: DEBUG nova.compute.claims [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 755.488103] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.703957] env[61986]: DEBUG nova.scheduler.client.report [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.718578] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 755.743783] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.744052] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.744210] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.744427] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.744586] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.744733] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.744929] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.745094] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.745258] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.745415] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.745582] env[61986]: DEBUG nova.virt.hardware [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.746461] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca43c741-0201-4ecd-a604-a160a7f896b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.755070] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263c45da-850d-4025-a8f0-f47604f67cfc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.768410] env[61986]: ERROR nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Traceback (most recent call last): [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] yield resources [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self.driver.spawn(context, instance, image_meta, [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] vm_ref = self.build_virtual_machine(instance, [ 755.768410] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] for vif in network_info: [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] return self._sync_wrapper(fn, *args, **kwargs) [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self.wait() [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self[:] = self._gt.wait() [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] return self._exit_event.wait() [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.768802] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] current.throw(*self._exc) [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] result = function(*args, **kwargs) [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] return func(*args, **kwargs) [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] raise e [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] nwinfo = self.network_api.allocate_for_instance( [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] created_port_ids = self._update_ports_for_instance( [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] with excutils.save_and_reraise_exception(): [ 755.769146] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self.force_reraise() [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] raise self.value [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] updated_port = self._update_port( [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] _ensure_no_port_binding_failure(port) [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] raise exception.PortBindingFailed(port_id=port['id']) [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 755.769481] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] [ 755.769481] env[61986]: INFO nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Terminating instance [ 755.770652] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.770806] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquired lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.770966] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.826894] env[61986]: DEBUG oslo_concurrency.lockutils [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] Releasing lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.827177] env[61986]: DEBUG nova.compute.manager [req-970a3822-0dda-449d-9c10-3d8318e1b32e req-99febcc9-53c5-49dc-95ff-055634082966 service nova] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Received event network-vif-deleted-925120df-b890-4643-91ee-74efd811ab1c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 756.208575] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.209159] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.211878] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.422s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.292220] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.374020] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.716877] env[61986]: DEBUG nova.compute.utils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.721500] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.721677] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 756.749635] env[61986]: DEBUG nova.compute.manager [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Received event network-changed-b813b6da-fc0a-4458-b1fb-9e8992b72c92 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 756.749822] env[61986]: DEBUG nova.compute.manager [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Refreshing instance network info cache due to event network-changed-b813b6da-fc0a-4458-b1fb-9e8992b72c92. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 756.750013] env[61986]: DEBUG oslo_concurrency.lockutils [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] Acquiring lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.767898] env[61986]: DEBUG nova.policy [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aeee0c50616642739124854e6c5d7cb0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '370e1ac7e7e84b48b60cf451dba0ddcb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 756.876700] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Releasing lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.877129] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.877324] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 756.877630] env[61986]: DEBUG oslo_concurrency.lockutils [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] Acquired lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.877794] env[61986]: DEBUG nova.network.neutron [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Refreshing network info cache for port b813b6da-fc0a-4458-b1fb-9e8992b72c92 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 756.878823] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed8f6d64-b094-4ad1-965b-f381b9f42c4b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.888790] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63918e8-064a-45a6-a1e0-c1ee53098d8c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.912756] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 77f4d9bd-c758-44c3-a17c-1efa8c395eb1 could not be found. [ 756.913019] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 756.913165] env[61986]: INFO nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 756.913404] env[61986]: DEBUG oslo.service.loopingcall [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.915684] env[61986]: DEBUG nova.compute.manager [-] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.915817] env[61986]: DEBUG nova.network.neutron [-] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.931884] env[61986]: DEBUG nova.network.neutron [-] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.068912] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Successfully created port: a7dc29a6-08a7-4471-9ae7-144a0d936ff7 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.128159] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548a7598-6859-46c0-b4c9-36e8053bf5f0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.135938] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4652e4f0-c116-4e19-840e-98c1ab998ef2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.165988] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578d673e-a399-4d62-9c3c-2128319feedb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.173452] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2c35f7-1565-41f9-8b9e-9bf9842da0da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.186670] env[61986]: DEBUG nova.compute.provider_tree [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.224813] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.399015] env[61986]: DEBUG nova.network.neutron [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.435839] env[61986]: DEBUG nova.network.neutron [-] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.497638] env[61986]: DEBUG nova.network.neutron [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.690289] env[61986]: DEBUG nova.scheduler.client.report [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.906589] env[61986]: ERROR nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 757.906589] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.906589] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.906589] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.906589] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.906589] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.906589] env[61986]: ERROR nova.compute.manager raise self.value [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.906589] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.906589] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.906589] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.907111] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.907111] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.907111] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 757.907111] env[61986]: ERROR nova.compute.manager [ 757.907111] env[61986]: Traceback (most recent call last): [ 757.907111] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.907111] env[61986]: listener.cb(fileno) [ 757.907111] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.907111] env[61986]: result = function(*args, **kwargs) [ 757.907111] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.907111] env[61986]: return func(*args, **kwargs) [ 757.907111] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.907111] env[61986]: raise e [ 757.907111] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.907111] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 757.907111] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.907111] env[61986]: created_port_ids = self._update_ports_for_instance( [ 757.907111] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.907111] env[61986]: with excutils.save_and_reraise_exception(): [ 757.907111] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.907111] env[61986]: self.force_reraise() [ 757.907111] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.907111] env[61986]: raise self.value [ 757.907111] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.907111] env[61986]: updated_port = self._update_port( [ 757.907111] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.907111] env[61986]: _ensure_no_port_binding_failure(port) [ 757.907111] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.907111] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.908059] env[61986]: nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 757.908059] env[61986]: Removing descriptor: 17 [ 757.937931] env[61986]: INFO nova.compute.manager [-] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Took 1.02 seconds to deallocate network for instance. [ 757.940413] env[61986]: DEBUG nova.compute.claims [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 757.940647] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.000648] env[61986]: DEBUG oslo_concurrency.lockutils [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] Releasing lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.001086] env[61986]: DEBUG nova.compute.manager [req-4844448c-1b3e-4ec2-9e8c-5046e81711c1 req-6713202d-5d57-4266-8f99-782237e030b1 service nova] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Received event network-vif-deleted-b813b6da-fc0a-4458-b1fb-9e8992b72c92 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 758.195204] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.195874] env[61986]: ERROR nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Traceback (most recent call last): [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self.driver.spawn(context, instance, image_meta, [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] vm_ref = self.build_virtual_machine(instance, [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.195874] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] for vif in network_info: [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return self._sync_wrapper(fn, *args, **kwargs) [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self.wait() [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self[:] = self._gt.wait() [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return self._exit_event.wait() [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] result = hub.switch() [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.196222] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return self.greenlet.switch() [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] result = function(*args, **kwargs) [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] return func(*args, **kwargs) [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] raise e [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] nwinfo = self.network_api.allocate_for_instance( [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] created_port_ids = self._update_ports_for_instance( [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] with excutils.save_and_reraise_exception(): [ 758.196638] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] self.force_reraise() [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] raise self.value [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] updated_port = self._update_port( [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] _ensure_no_port_binding_failure(port) [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] raise exception.PortBindingFailed(port_id=port['id']) [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] nova.exception.PortBindingFailed: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. [ 758.197036] env[61986]: ERROR nova.compute.manager [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] [ 758.197391] env[61986]: DEBUG nova.compute.utils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 758.197885] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.899s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.200681] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Build of instance fec98558-ff3a-4d65-a650-42b6c3ab9e18 was re-scheduled: Binding failed for port 5634e486-a2ab-4523-992a-fd0379b1f0db, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 758.201115] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 758.201332] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.201476] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquired lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.201629] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.232186] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.256247] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.256533] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.256698] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.256881] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.257035] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.257182] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.257381] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.257560] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.257732] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.258022] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.258344] env[61986]: DEBUG nova.virt.hardware [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.258967] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20e8cfe-b2fc-4f7e-99d7-5bf2cacbe249 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.266851] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca76a7ce-fb04-4dd1-a8b7-712cbf4fd23f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.281107] env[61986]: ERROR nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] Traceback (most recent call last): [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] yield resources [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self.driver.spawn(context, instance, image_meta, [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] vm_ref = self.build_virtual_machine(instance, [ 758.281107] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] for vif in network_info: [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] return self._sync_wrapper(fn, *args, **kwargs) [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self.wait() [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self[:] = self._gt.wait() [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] return self._exit_event.wait() [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.281600] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] current.throw(*self._exc) [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] result = function(*args, **kwargs) [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] return func(*args, **kwargs) [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] raise e [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] nwinfo = self.network_api.allocate_for_instance( [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] created_port_ids = self._update_ports_for_instance( [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] with excutils.save_and_reraise_exception(): [ 758.281981] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self.force_reraise() [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] raise self.value [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] updated_port = self._update_port( [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] _ensure_no_port_binding_failure(port) [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] raise exception.PortBindingFailed(port_id=port['id']) [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 758.282368] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] [ 758.282368] env[61986]: INFO nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Terminating instance [ 758.283517] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquiring lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.283679] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquired lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.283841] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.724479] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.785305] env[61986]: DEBUG nova.compute.manager [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] [instance: 55102fef-de7c-4505-a192-39568e08c096] Received event network-changed-a7dc29a6-08a7-4471-9ae7-144a0d936ff7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 758.785563] env[61986]: DEBUG nova.compute.manager [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] [instance: 55102fef-de7c-4505-a192-39568e08c096] Refreshing instance network info cache due to event network-changed-a7dc29a6-08a7-4471-9ae7-144a0d936ff7. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 758.785683] env[61986]: DEBUG oslo_concurrency.lockutils [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] Acquiring lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.801444] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.813517] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.869134] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.030896] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360eb295-ada8-445d-976c-07c900ad1e87 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.038180] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006f6966-7459-4b87-8021-57d9368691ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.069068] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37fa4b95-2e2f-4ebd-bcf4-cb4efb18c786 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.076091] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c61efa-a478-4ff5-8a48-b2e5a7bc9649 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.088974] env[61986]: DEBUG nova.compute.provider_tree [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.315692] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Releasing lock "refresh_cache-fec98558-ff3a-4d65-a650-42b6c3ab9e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.315990] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 759.316255] env[61986]: DEBUG nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.316523] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.337505] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.371154] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Releasing lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.371565] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.371756] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 759.372062] env[61986]: DEBUG oslo_concurrency.lockutils [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] Acquired lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.372233] env[61986]: DEBUG nova.network.neutron [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] [instance: 55102fef-de7c-4505-a192-39568e08c096] Refreshing network info cache for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.373209] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b236cbac-76a3-4344-ab21-b32426d37f67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.383260] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb998977-289c-464c-afe6-75e7fd635c35 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.404262] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 55102fef-de7c-4505-a192-39568e08c096 could not be found. [ 759.404468] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 759.404646] env[61986]: INFO nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.404879] env[61986]: DEBUG oslo.service.loopingcall [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.405128] env[61986]: DEBUG nova.compute.manager [-] [instance: 55102fef-de7c-4505-a192-39568e08c096] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.405230] env[61986]: DEBUG nova.network.neutron [-] [instance: 55102fef-de7c-4505-a192-39568e08c096] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.430752] env[61986]: DEBUG nova.network.neutron [-] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.592285] env[61986]: DEBUG nova.scheduler.client.report [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.840968] env[61986]: DEBUG nova.network.neutron [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.891375] env[61986]: DEBUG nova.network.neutron [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.933607] env[61986]: DEBUG nova.network.neutron [-] [instance: 55102fef-de7c-4505-a192-39568e08c096] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.999841] env[61986]: DEBUG nova.network.neutron [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] [instance: 55102fef-de7c-4505-a192-39568e08c096] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.097422] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.899s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.098126] env[61986]: ERROR nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Traceback (most recent call last): [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self.driver.spawn(context, instance, image_meta, [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] vm_ref = self.build_virtual_machine(instance, [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.098126] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] for vif in network_info: [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return self._sync_wrapper(fn, *args, **kwargs) [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self.wait() [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self[:] = self._gt.wait() [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return self._exit_event.wait() [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] result = hub.switch() [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.098474] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return self.greenlet.switch() [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] result = function(*args, **kwargs) [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] return func(*args, **kwargs) [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] raise e [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] nwinfo = self.network_api.allocate_for_instance( [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] created_port_ids = self._update_ports_for_instance( [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] with excutils.save_and_reraise_exception(): [ 760.098861] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] self.force_reraise() [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] raise self.value [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] updated_port = self._update_port( [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] _ensure_no_port_binding_failure(port) [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] raise exception.PortBindingFailed(port_id=port['id']) [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] nova.exception.PortBindingFailed: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. [ 760.099247] env[61986]: ERROR nova.compute.manager [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] [ 760.099562] env[61986]: DEBUG nova.compute.utils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.101855] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.986s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.102513] env[61986]: INFO nova.compute.claims [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.105102] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Build of instance 18f91633-5899-4f6d-addb-03c7ec39bc2c was re-scheduled: Binding failed for port 1e949011-79f1-4aae-99bf-a73b87b96104, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 760.105534] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 760.105754] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.105894] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquired lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.106060] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.343714] env[61986]: INFO nova.compute.manager [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: fec98558-ff3a-4d65-a650-42b6c3ab9e18] Took 1.03 seconds to deallocate network for instance. [ 760.437053] env[61986]: INFO nova.compute.manager [-] [instance: 55102fef-de7c-4505-a192-39568e08c096] Took 1.03 seconds to deallocate network for instance. [ 760.438622] env[61986]: DEBUG nova.compute.claims [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 760.438800] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.502827] env[61986]: DEBUG oslo_concurrency.lockutils [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] Releasing lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.503107] env[61986]: DEBUG nova.compute.manager [req-bb00a06c-bfab-496e-84c3-abf3db060c01 req-6fd34497-5dd2-42b7-8dec-4a307082dd70 service nova] [instance: 55102fef-de7c-4505-a192-39568e08c096] Received event network-vif-deleted-a7dc29a6-08a7-4471-9ae7-144a0d936ff7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 760.624146] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.723414] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.226191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Releasing lock "refresh_cache-18f91633-5899-4f6d-addb-03c7ec39bc2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.226434] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.226687] env[61986]: DEBUG nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.226859] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.249358] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.375998] env[61986]: INFO nova.scheduler.client.report [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Deleted allocations for instance fec98558-ff3a-4d65-a650-42b6c3ab9e18 [ 761.441881] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24601a5-6840-4fa3-b3c6-23d00efe4a47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.448771] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594f1299-9f8c-4018-9964-512eb5877556 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.478283] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591a2bde-c560-45a6-b545-28782bd42a9d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.485462] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc03038-302b-456b-a38b-ae0be6553a76 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.498788] env[61986]: DEBUG nova.compute.provider_tree [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.752590] env[61986]: DEBUG nova.network.neutron [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.887836] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1c8f43b-58de-49aa-a5cb-e07cc9539ad6 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "fec98558-ff3a-4d65-a650-42b6c3ab9e18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.133s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.001442] env[61986]: DEBUG nova.scheduler.client.report [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.255950] env[61986]: INFO nova.compute.manager [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 18f91633-5899-4f6d-addb-03c7ec39bc2c] Took 1.03 seconds to deallocate network for instance. [ 762.390427] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.507368] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.508024] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.510304] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.636s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.912216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.018046] env[61986]: DEBUG nova.compute.utils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.019619] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.019758] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 763.071078] env[61986]: DEBUG nova.policy [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb4b3a6fb1c7426796f38978b04940d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdab411a34424bc185f8a5981809437c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 763.281934] env[61986]: INFO nova.scheduler.client.report [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Deleted allocations for instance 18f91633-5899-4f6d-addb-03c7ec39bc2c [ 763.362195] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e88074-39c8-461d-897b-ab41240d8ee1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.370023] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb9287e-d09c-4f14-adaa-4ade245c97ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.402810] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7711bdd3-71f0-4fc7-99ab-6815267955c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.409992] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c070005a-7422-4095-9e28-75a03b922f18 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.422972] env[61986]: DEBUG nova.compute.provider_tree [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.428486] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Successfully created port: 9eb658ee-3092-4b43-af55-3e7718b0b616 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.522417] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.794758] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d2f7794f-8f22-486e-b518-1fc471f34b08 tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "18f91633-5899-4f6d-addb-03c7ec39bc2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.394s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.925949] env[61986]: DEBUG nova.scheduler.client.report [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.297140] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.322413] env[61986]: DEBUG nova.compute.manager [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Received event network-changed-9eb658ee-3092-4b43-af55-3e7718b0b616 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 764.322413] env[61986]: DEBUG nova.compute.manager [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Refreshing instance network info cache due to event network-changed-9eb658ee-3092-4b43-af55-3e7718b0b616. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 764.322413] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] Acquiring lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.322657] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] Acquired lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.322705] env[61986]: DEBUG nova.network.neutron [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Refreshing network info cache for port 9eb658ee-3092-4b43-af55-3e7718b0b616 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.431354] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.431994] env[61986]: ERROR nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Traceback (most recent call last): [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self.driver.spawn(context, instance, image_meta, [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] vm_ref = self.build_virtual_machine(instance, [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.431994] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] for vif in network_info: [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] return self._sync_wrapper(fn, *args, **kwargs) [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self.wait() [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self[:] = self._gt.wait() [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] return self._exit_event.wait() [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] current.throw(*self._exc) [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.432459] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] result = function(*args, **kwargs) [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] return func(*args, **kwargs) [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] raise e [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] nwinfo = self.network_api.allocate_for_instance( [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] created_port_ids = self._update_ports_for_instance( [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] with excutils.save_and_reraise_exception(): [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] self.force_reraise() [ 764.432793] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] raise self.value [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] updated_port = self._update_port( [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] _ensure_no_port_binding_failure(port) [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] raise exception.PortBindingFailed(port_id=port['id']) [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] nova.exception.PortBindingFailed: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. [ 764.433189] env[61986]: ERROR nova.compute.manager [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] [ 764.433189] env[61986]: DEBUG nova.compute.utils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 764.434253] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.911s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.437227] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Build of instance 213957b6-77a4-48fe-8bf5-f03bb67510c6 was re-scheduled: Binding failed for port 8d610236-808e-469e-b304-e781a4cc5fa2, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 764.437715] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 764.437944] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquiring lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.438098] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Acquired lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.438257] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.532418] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.559982] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.560318] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.560782] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.560782] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.560906] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.561446] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.561446] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.561595] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.561816] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.562058] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.562357] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.563406] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5ec419-0686-4c0c-9f42-1a404af1990b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.571652] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8910e08-7edb-477a-a732-467878a53336 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.654694] env[61986]: ERROR nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 764.654694] env[61986]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.654694] env[61986]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 764.654694] env[61986]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 764.654694] env[61986]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.654694] env[61986]: ERROR nova.compute.manager self.force_reraise() [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.654694] env[61986]: ERROR nova.compute.manager raise self.value [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 764.654694] env[61986]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.654694] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.654694] env[61986]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.655200] env[61986]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.655200] env[61986]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.655200] env[61986]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 764.655200] env[61986]: ERROR nova.compute.manager [ 764.655200] env[61986]: Traceback (most recent call last): [ 764.655200] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.655200] env[61986]: listener.cb(fileno) [ 764.655200] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.655200] env[61986]: result = function(*args, **kwargs) [ 764.655200] env[61986]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 764.655200] env[61986]: return func(*args, **kwargs) [ 764.655200] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.655200] env[61986]: raise e [ 764.655200] env[61986]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.655200] env[61986]: nwinfo = self.network_api.allocate_for_instance( [ 764.655200] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 764.655200] env[61986]: created_port_ids = self._update_ports_for_instance( [ 764.655200] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 764.655200] env[61986]: with excutils.save_and_reraise_exception(): [ 764.655200] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.655200] env[61986]: self.force_reraise() [ 764.655200] env[61986]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.655200] env[61986]: raise self.value [ 764.655200] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 764.655200] env[61986]: updated_port = self._update_port( [ 764.655200] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.655200] env[61986]: _ensure_no_port_binding_failure(port) [ 764.655200] env[61986]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.655200] env[61986]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.655908] env[61986]: nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 764.655908] env[61986]: Removing descriptor: 14 [ 764.655908] env[61986]: ERROR nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Traceback (most recent call last): [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] yield resources [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self.driver.spawn(context, instance, image_meta, [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.655908] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] vm_ref = self.build_virtual_machine(instance, [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] for vif in network_info: [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return self._sync_wrapper(fn, *args, **kwargs) [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self.wait() [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self[:] = self._gt.wait() [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return self._exit_event.wait() [ 764.656216] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] result = hub.switch() [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return self.greenlet.switch() [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] result = function(*args, **kwargs) [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return func(*args, **kwargs) [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] raise e [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] nwinfo = self.network_api.allocate_for_instance( [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 764.656557] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] created_port_ids = self._update_ports_for_instance( [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] with excutils.save_and_reraise_exception(): [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self.force_reraise() [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] raise self.value [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] updated_port = self._update_port( [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] _ensure_no_port_binding_failure(port) [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.656958] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] raise exception.PortBindingFailed(port_id=port['id']) [ 764.657290] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 764.657290] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] [ 764.657290] env[61986]: INFO nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Terminating instance [ 764.658551] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.819927] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.840901] env[61986]: DEBUG nova.network.neutron [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.985919] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.062183] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.112309] env[61986]: DEBUG nova.network.neutron [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.292930] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917ea15b-81d2-41c5-b1da-8ea5c6bc7633 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.300319] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d9bf89-39cd-45e3-976c-e3b6b4138ef1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.329651] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5267066-bb0a-4a40-9a70-cc3d71518d90 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.336698] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9262fd27-58e8-40ba-b026-c662028c537c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.349606] env[61986]: DEBUG nova.compute.provider_tree [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.563690] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Releasing lock "refresh_cache-213957b6-77a4-48fe-8bf5-f03bb67510c6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.563961] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 765.564158] env[61986]: DEBUG nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.564362] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 765.585247] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.617014] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e782c33-5a2f-4990-86fe-b8eea4b22a3c req-953e1ce9-9df7-4ab6-9c20-8de45c21770a service nova] Releasing lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.617420] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.617621] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.852209] env[61986]: DEBUG nova.scheduler.client.report [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.087410] env[61986]: DEBUG nova.network.neutron [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.140230] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.278810] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.353084] env[61986]: DEBUG nova.compute.manager [req-a7eec42b-7e3e-4387-ac63-fb794438b46c req-b8341368-6460-4940-acae-04585afaee0b service nova] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Received event network-vif-deleted-9eb658ee-3092-4b43-af55-3e7718b0b616 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 766.358650] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.924s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.359237] env[61986]: ERROR nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Traceback (most recent call last): [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self.driver.spawn(context, instance, image_meta, [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] vm_ref = self.build_virtual_machine(instance, [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.359237] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] for vif in network_info: [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] return self._sync_wrapper(fn, *args, **kwargs) [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self.wait() [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self[:] = self._gt.wait() [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] return self._exit_event.wait() [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] current.throw(*self._exc) [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.359594] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] result = function(*args, **kwargs) [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] return func(*args, **kwargs) [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] raise e [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] nwinfo = self.network_api.allocate_for_instance( [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] created_port_ids = self._update_ports_for_instance( [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] with excutils.save_and_reraise_exception(): [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] self.force_reraise() [ 766.359966] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] raise self.value [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] updated_port = self._update_port( [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] _ensure_no_port_binding_failure(port) [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] raise exception.PortBindingFailed(port_id=port['id']) [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] nova.exception.PortBindingFailed: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. [ 766.360345] env[61986]: ERROR nova.compute.manager [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] [ 766.360345] env[61986]: DEBUG nova.compute.utils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.361135] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.391s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.362468] env[61986]: INFO nova.compute.claims [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.365116] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Build of instance dd239b21-fdb2-48fb-acf4-18242a98cd1f was re-scheduled: Binding failed for port b307ac55-fe5e-4042-84a3-f24a463c1434, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 766.366411] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 766.366411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquiring lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.366411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Acquired lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.366411] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 766.594403] env[61986]: INFO nova.compute.manager [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] [instance: 213957b6-77a4-48fe-8bf5-f03bb67510c6] Took 1.03 seconds to deallocate network for instance. [ 766.782239] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.782679] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 766.782864] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 766.783215] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94029827-c2d4-476d-a70b-4b6120b85e9d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.792437] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a1cf4e-3df1-4a6e-beb4-b08597f7fbb1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.813732] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8bc12f2b-a632-4ffe-9d55-1a838e94931b could not be found. [ 766.813732] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 766.813927] env[61986]: INFO nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 766.814165] env[61986]: DEBUG oslo.service.loopingcall [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.814389] env[61986]: DEBUG nova.compute.manager [-] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.814484] env[61986]: DEBUG nova.network.neutron [-] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.829165] env[61986]: DEBUG nova.network.neutron [-] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.884842] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.951382] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.331797] env[61986]: DEBUG nova.network.neutron [-] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.455032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Releasing lock "refresh_cache-dd239b21-fdb2-48fb-acf4-18242a98cd1f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.455289] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 767.455465] env[61986]: DEBUG nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.455629] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 767.485650] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.642335] env[61986]: INFO nova.scheduler.client.report [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Deleted allocations for instance 213957b6-77a4-48fe-8bf5-f03bb67510c6 [ 767.685987] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa2071b-c89e-455c-9b81-94006e0eb858 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.693823] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8ffe42-7128-45d6-a176-a1cd5ccf3c4d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.723626] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4398868-98fc-4eec-9d94-e6ccf094a84d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.731022] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317cda74-2d5f-44c6-b6ac-3b493ce5792c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.744291] env[61986]: DEBUG nova.compute.provider_tree [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.834420] env[61986]: INFO nova.compute.manager [-] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Took 1.02 seconds to deallocate network for instance. [ 767.837525] env[61986]: DEBUG nova.compute.claims [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Aborting claim: {{(pid=61986) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 767.837739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.991213] env[61986]: DEBUG nova.network.neutron [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.152597] env[61986]: DEBUG oslo_concurrency.lockutils [None req-20277bde-1024-4dca-bddc-781a431ff95c tempest-ListServerFiltersTestJSON-1268752299 tempest-ListServerFiltersTestJSON-1268752299-project-member] Lock "213957b6-77a4-48fe-8bf5-f03bb67510c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.455s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.247324] env[61986]: DEBUG nova.scheduler.client.report [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.494230] env[61986]: INFO nova.compute.manager [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] [instance: dd239b21-fdb2-48fb-acf4-18242a98cd1f] Took 1.04 seconds to deallocate network for instance. [ 768.655498] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.754470] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.754470] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.756846] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.757s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.187282] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.262175] env[61986]: DEBUG nova.compute.utils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.267524] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.267763] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 769.359031] env[61986]: DEBUG nova.policy [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb4b3a6fb1c7426796f38978b04940d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdab411a34424bc185f8a5981809437c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 769.536179] env[61986]: INFO nova.scheduler.client.report [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Deleted allocations for instance dd239b21-fdb2-48fb-acf4-18242a98cd1f [ 769.627395] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563b90af-3621-402c-a4d8-86540bf88182 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.635684] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03846781-79c1-49e4-b5fc-95b7bfdb7fff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.670218] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896b6c12-6ad0-4ecf-ab49-48991893e111 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.679660] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef21d412-9721-45ca-a4d0-576c6a6e4b0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.695699] env[61986]: DEBUG nova.compute.provider_tree [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.747493] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Successfully created port: 91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.768204] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.045351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3f67cb25-3984-4582-a7c9-ecc974720a51 tempest-ServersTestManualDisk-1919927151 tempest-ServersTestManualDisk-1919927151-project-member] Lock "dd239b21-fdb2-48fb-acf4-18242a98cd1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.613s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.198712] env[61986]: DEBUG nova.scheduler.client.report [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.548999] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.708602] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.709269] env[61986]: ERROR nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Traceback (most recent call last): [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self.driver.spawn(context, instance, image_meta, [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] vm_ref = self.build_virtual_machine(instance, [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.709269] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] for vif in network_info: [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] return self._sync_wrapper(fn, *args, **kwargs) [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self.wait() [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self[:] = self._gt.wait() [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] return self._exit_event.wait() [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] current.throw(*self._exc) [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.709571] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] result = function(*args, **kwargs) [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] return func(*args, **kwargs) [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] raise e [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] nwinfo = self.network_api.allocate_for_instance( [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] created_port_ids = self._update_ports_for_instance( [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] with excutils.save_and_reraise_exception(): [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] self.force_reraise() [ 770.709895] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] raise self.value [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] updated_port = self._update_port( [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] _ensure_no_port_binding_failure(port) [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] raise exception.PortBindingFailed(port_id=port['id']) [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] nova.exception.PortBindingFailed: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. [ 770.710224] env[61986]: ERROR nova.compute.manager [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] [ 770.710224] env[61986]: DEBUG nova.compute.utils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 770.711426] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.223s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.717410] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Build of instance 7e5fadf3-dd41-402c-9878-2d11295203ac was re-scheduled: Binding failed for port 5b75aa7a-be06-41e3-a4a3-42e9e5b541d3, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 770.717410] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 770.717410] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquiring lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.717410] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Acquired lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.717410] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.780376] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.813548] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.814037] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.814424] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.814771] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.815076] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.815430] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.815768] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.816066] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.816366] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.816682] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.817065] env[61986]: DEBUG nova.virt.hardware [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.818060] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fc76f4-92ec-4906-be1c-ab4df0151ebb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.826468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4bf5a4-2309-45be-8589-129ba469dc87 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.075670] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.246298] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.394513] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.596745] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed1814f-d3ce-4720-9707-2dd7784f52ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.608975] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a746e860-d47c-4478-8a5a-487289647ad9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.644791] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0e450e-a1ff-4d2a-b156-80e017757b6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.657164] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c03319-479b-460e-9404-3b4b28835aef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.673086] env[61986]: DEBUG nova.compute.provider_tree [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.889909] env[61986]: DEBUG nova.compute.manager [req-e519423d-6af1-4321-b21c-52e1b322e379 req-82616d06-edb5-4781-a196-79c0d2c8a4d4 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Received event network-vif-plugged-91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.890474] env[61986]: DEBUG oslo_concurrency.lockutils [req-e519423d-6af1-4321-b21c-52e1b322e379 req-82616d06-edb5-4781-a196-79c0d2c8a4d4 service nova] Acquiring lock "2377b482-52e7-4054-8d5b-8f00587c3371-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.890643] env[61986]: DEBUG oslo_concurrency.lockutils [req-e519423d-6af1-4321-b21c-52e1b322e379 req-82616d06-edb5-4781-a196-79c0d2c8a4d4 service nova] Lock "2377b482-52e7-4054-8d5b-8f00587c3371-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.890702] env[61986]: DEBUG oslo_concurrency.lockutils [req-e519423d-6af1-4321-b21c-52e1b322e379 req-82616d06-edb5-4781-a196-79c0d2c8a4d4 service nova] Lock "2377b482-52e7-4054-8d5b-8f00587c3371-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.891029] env[61986]: DEBUG nova.compute.manager [req-e519423d-6af1-4321-b21c-52e1b322e379 req-82616d06-edb5-4781-a196-79c0d2c8a4d4 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] No waiting events found dispatching network-vif-plugged-91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 771.891433] env[61986]: WARNING nova.compute.manager [req-e519423d-6af1-4321-b21c-52e1b322e379 req-82616d06-edb5-4781-a196-79c0d2c8a4d4 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Received unexpected event network-vif-plugged-91d1d769-094d-4ae2-84a2-868652bd2b82 for instance with vm_state building and task_state spawning. [ 771.896559] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Releasing lock "refresh_cache-7e5fadf3-dd41-402c-9878-2d11295203ac" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.896780] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 771.896960] env[61986]: DEBUG nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.897136] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 771.923056] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.099306] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Successfully updated port: 91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 772.180255] env[61986]: DEBUG nova.scheduler.client.report [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.425582] env[61986]: DEBUG nova.network.neutron [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.602447] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.602571] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.602719] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.685388] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.686324] env[61986]: ERROR nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Traceback (most recent call last): [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self.driver.spawn(context, instance, image_meta, [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] vm_ref = self.build_virtual_machine(instance, [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.686324] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] for vif in network_info: [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] return self._sync_wrapper(fn, *args, **kwargs) [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self.wait() [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self[:] = self._gt.wait() [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] return self._exit_event.wait() [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] current.throw(*self._exc) [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.686721] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] result = function(*args, **kwargs) [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] return func(*args, **kwargs) [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] raise e [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] nwinfo = self.network_api.allocate_for_instance( [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] created_port_ids = self._update_ports_for_instance( [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] with excutils.save_and_reraise_exception(): [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] self.force_reraise() [ 772.687078] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] raise self.value [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] updated_port = self._update_port( [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] _ensure_no_port_binding_failure(port) [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] raise exception.PortBindingFailed(port_id=port['id']) [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] nova.exception.PortBindingFailed: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. [ 772.687479] env[61986]: ERROR nova.compute.manager [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] [ 772.688732] env[61986]: DEBUG nova.compute.utils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.691471] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.751s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.694246] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Build of instance 198f36d0-8be4-4885-9c0b-e85d6a0cbb22 was re-scheduled: Binding failed for port 925120df-b890-4643-91ee-74efd811ab1c, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 772.697040] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 772.697040] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.697040] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquired lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.697040] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.930600] env[61986]: INFO nova.compute.manager [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] [instance: 7e5fadf3-dd41-402c-9878-2d11295203ac] Took 1.03 seconds to deallocate network for instance. [ 773.151930] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.222712] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.393870] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updating instance_info_cache with network_info: [{"id": "91d1d769-094d-4ae2-84a2-868652bd2b82", "address": "fa:16:3e:3d:75:11", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91d1d769-09", "ovs_interfaceid": "91d1d769-094d-4ae2-84a2-868652bd2b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.395718] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.597864] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0d0789-57ae-407f-8a8a-95b4094b4fcb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.605669] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adff5fe-f30c-44c7-869f-591d8db34a68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.635692] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029e0607-f5d2-4c84-a15d-47c9b9cb0701 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.643288] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44044dc6-49c1-4fa1-bf93-d54a886d2c62 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.656928] env[61986]: DEBUG nova.compute.provider_tree [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.899610] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.899950] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Instance network_info: |[{"id": "91d1d769-094d-4ae2-84a2-868652bd2b82", "address": "fa:16:3e:3d:75:11", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91d1d769-09", "ovs_interfaceid": "91d1d769-094d-4ae2-84a2-868652bd2b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 773.902406] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:75:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91d1d769-094d-4ae2-84a2-868652bd2b82', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 773.912692] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Creating folder: Project (fdab411a34424bc185f8a5981809437c). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.914225] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Releasing lock "refresh_cache-198f36d0-8be4-4885-9c0b-e85d6a0cbb22" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.914225] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 773.914371] env[61986]: DEBUG nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.914564] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.916796] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3dc67ff-631c-4a72-89c2-319ed95aa6c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.929661] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Created folder: Project (fdab411a34424bc185f8a5981809437c) in parent group-v252271. [ 773.929661] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Creating folder: Instances. Parent ref: group-v252280. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.929856] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8fd01df3-fb47-458e-91fa-f1b2903ecf7f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.939067] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Created folder: Instances in parent group-v252280. [ 773.939896] env[61986]: DEBUG oslo.service.loopingcall [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.939896] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 773.939896] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d0af8608-1bb5-4f2f-b845-8cb62f7e8ba1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.956971] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.961617] env[61986]: DEBUG nova.compute.manager [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Received event network-changed-91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.962471] env[61986]: DEBUG nova.compute.manager [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Refreshing instance network info cache due to event network-changed-91d1d769-094d-4ae2-84a2-868652bd2b82. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 773.962471] env[61986]: DEBUG oslo_concurrency.lockutils [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] Acquiring lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.962471] env[61986]: DEBUG oslo_concurrency.lockutils [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] Acquired lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.962471] env[61986]: DEBUG nova.network.neutron [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Refreshing network info cache for port 91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.970410] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 773.970410] env[61986]: value = "task-1159538" [ 773.970410] env[61986]: _type = "Task" [ 773.970410] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.984610] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159538, 'name': CreateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.988596] env[61986]: INFO nova.scheduler.client.report [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Deleted allocations for instance 7e5fadf3-dd41-402c-9878-2d11295203ac [ 774.159934] env[61986]: DEBUG nova.scheduler.client.report [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.465115] env[61986]: DEBUG nova.network.neutron [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.480212] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159538, 'name': CreateVM_Task, 'duration_secs': 0.317803} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.480424] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 774.491113] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.491113] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.491113] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 774.491113] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6b04e26-58d5-4d4b-952b-c424f8be8e05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.494170] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 774.494170] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520254e3-14fc-64a8-9bfb-b36b6b5fed07" [ 774.494170] env[61986]: _type = "Task" [ 774.494170] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.505556] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d6e505a-253f-4ad2-93b5-286676006a74 tempest-ServersTestJSON-1613725255 tempest-ServersTestJSON-1613725255-project-member] Lock "7e5fadf3-dd41-402c-9878-2d11295203ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.254s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.506025] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520254e3-14fc-64a8-9bfb-b36b6b5fed07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.666211] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.666861] env[61986]: ERROR nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Traceback (most recent call last): [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self.driver.spawn(context, instance, image_meta, [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] vm_ref = self.build_virtual_machine(instance, [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.666861] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] for vif in network_info: [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] return self._sync_wrapper(fn, *args, **kwargs) [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self.wait() [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self[:] = self._gt.wait() [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] return self._exit_event.wait() [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] current.throw(*self._exc) [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.667216] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] result = function(*args, **kwargs) [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] return func(*args, **kwargs) [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] raise e [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] nwinfo = self.network_api.allocate_for_instance( [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] created_port_ids = self._update_ports_for_instance( [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] with excutils.save_and_reraise_exception(): [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] self.force_reraise() [ 774.667538] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] raise self.value [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] updated_port = self._update_port( [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] _ensure_no_port_binding_failure(port) [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] raise exception.PortBindingFailed(port_id=port['id']) [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] nova.exception.PortBindingFailed: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. [ 774.668023] env[61986]: ERROR nova.compute.manager [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] [ 774.668023] env[61986]: DEBUG nova.compute.utils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 774.668722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.230s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.671813] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Build of instance 77f4d9bd-c758-44c3-a17c-1efa8c395eb1 was re-scheduled: Binding failed for port b813b6da-fc0a-4458-b1fb-9e8992b72c92, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 774.672245] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 774.672464] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquiring lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.672606] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Acquired lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.672760] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.859126] env[61986]: DEBUG nova.network.neutron [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updated VIF entry in instance network info cache for port 91d1d769-094d-4ae2-84a2-868652bd2b82. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 774.859126] env[61986]: DEBUG nova.network.neutron [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updating instance_info_cache with network_info: [{"id": "91d1d769-094d-4ae2-84a2-868652bd2b82", "address": "fa:16:3e:3d:75:11", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91d1d769-09", "ovs_interfaceid": "91d1d769-094d-4ae2-84a2-868652bd2b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.969468] env[61986]: INFO nova.compute.manager [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: 198f36d0-8be4-4885-9c0b-e85d6a0cbb22] Took 1.05 seconds to deallocate network for instance. [ 775.011403] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.015126] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520254e3-14fc-64a8-9bfb-b36b6b5fed07, 'name': SearchDatastore_Task, 'duration_secs': 0.012122} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.016007] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.016627] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.016627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.016627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.018188] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.018188] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8b24e01-3bd6-4df7-aecf-789662538094 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.029223] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.029430] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 775.030170] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79426ff5-1737-499c-951f-f6e69960568f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.036016] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 775.036016] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d753fd-8df5-fb8b-32de-197c02436db0" [ 775.036016] env[61986]: _type = "Task" [ 775.036016] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.043985] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d753fd-8df5-fb8b-32de-197c02436db0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.197497] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.362820] env[61986]: DEBUG oslo_concurrency.lockutils [req-14ac42ad-4e83-4574-a509-b2dbd50ae6c6 req-8bb037f0-0a63-4790-8c45-2f2743498835 service nova] Releasing lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.444599] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.535276] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.548712] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d753fd-8df5-fb8b-32de-197c02436db0, 'name': SearchDatastore_Task, 'duration_secs': 0.009066} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.551187] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246c267c-3055-479f-bceb-fbd29d9aec19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.553605] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98576a3e-cd7e-4670-be97-968ffdce1a18 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.560950] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18bfe6f0-6c0d-4ed7-9f4c-1efc7e948487 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.564255] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 775.564255] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c0f935-48a3-d694-119d-9e71545c4b1b" [ 775.564255] env[61986]: _type = "Task" [ 775.564255] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.591154] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24653ac5-ea9b-43de-9f44-fedb90100fac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.596957] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c0f935-48a3-d694-119d-9e71545c4b1b, 'name': SearchDatastore_Task, 'duration_secs': 0.007975} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.597540] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.597806] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 2377b482-52e7-4054-8d5b-8f00587c3371/2377b482-52e7-4054-8d5b-8f00587c3371.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 775.598065] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69a24c81-2c1d-4364-98a0-81790a3e090f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.603154] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628db25c-ba60-4407-8a27-b1e857114e9b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.607664] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 775.607664] env[61986]: value = "task-1159539" [ 775.607664] env[61986]: _type = "Task" [ 775.607664] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.618138] env[61986]: DEBUG nova.compute.provider_tree [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.624019] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159539, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.946733] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Releasing lock "refresh_cache-77f4d9bd-c758-44c3-a17c-1efa8c395eb1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.946942] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 775.947154] env[61986]: DEBUG nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.947321] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 775.969064] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.011581] env[61986]: INFO nova.scheduler.client.report [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Deleted allocations for instance 198f36d0-8be4-4885-9c0b-e85d6a0cbb22 [ 776.119921] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159539, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483928} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.123242] env[61986]: DEBUG nova.scheduler.client.report [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.128039] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 2377b482-52e7-4054-8d5b-8f00587c3371/2377b482-52e7-4054-8d5b-8f00587c3371.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 776.128039] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.128039] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd301e76-01ca-4820-89c1-1cee8d312c7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.133627] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 776.133627] env[61986]: value = "task-1159540" [ 776.133627] env[61986]: _type = "Task" [ 776.133627] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.141970] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.475590] env[61986]: DEBUG nova.network.neutron [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.523745] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d8c8a2b8-28eb-4688-9a48-f8fe41ca9bd4 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "198f36d0-8be4-4885-9c0b-e85d6a0cbb22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.409s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.631768] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.632444] env[61986]: ERROR nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] Traceback (most recent call last): [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self.driver.spawn(context, instance, image_meta, [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] vm_ref = self.build_virtual_machine(instance, [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.632444] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] for vif in network_info: [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] return self._sync_wrapper(fn, *args, **kwargs) [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self.wait() [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self[:] = self._gt.wait() [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] return self._exit_event.wait() [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] current.throw(*self._exc) [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.632829] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] result = function(*args, **kwargs) [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] return func(*args, **kwargs) [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] raise e [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] nwinfo = self.network_api.allocate_for_instance( [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] created_port_ids = self._update_ports_for_instance( [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] with excutils.save_and_reraise_exception(): [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] self.force_reraise() [ 776.633216] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] raise self.value [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] updated_port = self._update_port( [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] _ensure_no_port_binding_failure(port) [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] raise exception.PortBindingFailed(port_id=port['id']) [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] nova.exception.PortBindingFailed: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. [ 776.633598] env[61986]: ERROR nova.compute.manager [instance: 55102fef-de7c-4505-a192-39568e08c096] [ 776.633598] env[61986]: DEBUG nova.compute.utils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.637342] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.723s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.637342] env[61986]: INFO nova.compute.claims [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.640150] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Build of instance 55102fef-de7c-4505-a192-39568e08c096 was re-scheduled: Binding failed for port a7dc29a6-08a7-4471-9ae7-144a0d936ff7, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 776.640578] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 776.640793] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquiring lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.640937] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Acquired lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.641104] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.653097] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.144741} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.653351] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.658134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991c8a80-31da-422a-b089-aecc17b8d3d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.696103] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 2377b482-52e7-4054-8d5b-8f00587c3371/2377b482-52e7-4054-8d5b-8f00587c3371.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.697127] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93af66f6-41af-493f-9af4-8a45f33dafae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.719493] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 776.719493] env[61986]: value = "task-1159541" [ 776.719493] env[61986]: _type = "Task" [ 776.719493] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.731024] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.978854] env[61986]: INFO nova.compute.manager [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] [instance: 77f4d9bd-c758-44c3-a17c-1efa8c395eb1] Took 1.03 seconds to deallocate network for instance. [ 777.027327] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.179782] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.233987] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159541, 'name': ReconfigVM_Task, 'duration_secs': 0.27931} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.234357] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 2377b482-52e7-4054-8d5b-8f00587c3371/2377b482-52e7-4054-8d5b-8f00587c3371.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.235167] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4afeee04-31a2-4953-9ae8-7276945115d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.242968] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 777.242968] env[61986]: value = "task-1159542" [ 777.242968] env[61986]: _type = "Task" [ 777.242968] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.254317] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159542, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.359243] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.552376] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.752805] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159542, 'name': Rename_Task, 'duration_secs': 0.147526} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.758940] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 777.760453] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6326ec6d-f3a4-42e0-9e4d-4809a0ff4859 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.770719] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 777.770719] env[61986]: value = "task-1159543" [ 777.770719] env[61986]: _type = "Task" [ 777.770719] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.781694] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.862427] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Releasing lock "refresh_cache-55102fef-de7c-4505-a192-39568e08c096" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.862674] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 777.862853] env[61986]: DEBUG nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.863028] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.882537] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.017824] env[61986]: INFO nova.scheduler.client.report [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Deleted allocations for instance 77f4d9bd-c758-44c3-a17c-1efa8c395eb1 [ 778.042759] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51136d1-5475-40db-b505-9f8897f64439 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.051574] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a937b7-cd5e-4709-ae1c-8cc3e746f175 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.084439] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc434843-bbca-4ca9-8c30-b74758e0b6f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.092164] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d971ac0-b62a-4115-beba-a5f7e37afa7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.107030] env[61986]: DEBUG nova.compute.provider_tree [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.286222] env[61986]: DEBUG oslo_vmware.api [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159543, 'name': PowerOnVM_Task, 'duration_secs': 0.464931} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.286543] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 778.286786] env[61986]: INFO nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Took 7.51 seconds to spawn the instance on the hypervisor. [ 778.286996] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.287933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56467a76-d192-4c54-b57b-fce9365bd93f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.387369] env[61986]: DEBUG nova.network.neutron [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.531479] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85746f83-304b-47cc-85eb-40c0c515d1b3 tempest-SecurityGroupsTestJSON-1167332968 tempest-SecurityGroupsTestJSON-1167332968-project-member] Lock "77f4d9bd-c758-44c3-a17c-1efa8c395eb1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.965s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.609976] env[61986]: DEBUG nova.scheduler.client.report [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.810716] env[61986]: INFO nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Took 27.85 seconds to build instance. [ 778.892525] env[61986]: INFO nova.compute.manager [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] [instance: 55102fef-de7c-4505-a192-39568e08c096] Took 1.03 seconds to deallocate network for instance. [ 779.037252] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.119141] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.119564] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.126235] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.302s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.127792] env[61986]: INFO nova.compute.claims [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.315282] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "2377b482-52e7-4054-8d5b-8f00587c3371" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.200s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.564043] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.645743] env[61986]: DEBUG nova.compute.utils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.649437] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.649942] env[61986]: DEBUG nova.network.neutron [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 779.715030] env[61986]: DEBUG nova.policy [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4675c2084b6463bb47c93865f5cec46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f746f57f466f480c958d8b15bd2ce8fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 779.820696] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "021a27ec-cee0-454d-8daf-e6a82bd9330c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.820696] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.820696] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.936882] env[61986]: INFO nova.scheduler.client.report [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Deleted allocations for instance 55102fef-de7c-4505-a192-39568e08c096 [ 780.035945] env[61986]: DEBUG nova.network.neutron [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Successfully created port: 038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.155983] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.344216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.418388] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.418388] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.450124] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3fb929c1-b730-4417-ac37-911cd11a1262 tempest-AttachInterfacesUnderV243Test-1774520345 tempest-AttachInterfacesUnderV243Test-1774520345-project-member] Lock "55102fef-de7c-4505-a192-39568e08c096" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.910s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.529954] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5c0669-1721-43b3-8793-d0787c2d6b1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.535600] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a03966-dca7-4ba7-873f-77f4b4ff856c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.573748] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c855bd-dfc8-4903-af70-adaf024b2d80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.582330] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3947b3ee-855c-4681-976b-80d4f0103f32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.597905] env[61986]: DEBUG nova.compute.provider_tree [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.953964] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.100926] env[61986]: DEBUG nova.scheduler.client.report [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.167913] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.198028] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.198337] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.198502] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.198688] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.198832] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.198997] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.199259] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.199422] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.199592] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.199769] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.199942] env[61986]: DEBUG nova.virt.hardware [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.200831] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180f73c4-8fe8-46f6-bfbf-43e62b27f0fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.209256] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622d7c29-af5d-490a-b5b8-249da7157c43 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.486501] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.610037] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.610037] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 781.612226] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.774s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.809424] env[61986]: DEBUG nova.network.neutron [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Successfully updated port: 038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.068693] env[61986]: DEBUG nova.compute.manager [req-a15f3c47-c8c6-4ec2-adbd-0e9bfdc41694 req-83be40a7-8da0-4a9a-a8ef-45d3849d910f service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Received event network-vif-plugged-038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.068851] env[61986]: DEBUG oslo_concurrency.lockutils [req-a15f3c47-c8c6-4ec2-adbd-0e9bfdc41694 req-83be40a7-8da0-4a9a-a8ef-45d3849d910f service nova] Acquiring lock "aa12919e-3891-4da9-a280-4155137864e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.069075] env[61986]: DEBUG oslo_concurrency.lockutils [req-a15f3c47-c8c6-4ec2-adbd-0e9bfdc41694 req-83be40a7-8da0-4a9a-a8ef-45d3849d910f service nova] Lock "aa12919e-3891-4da9-a280-4155137864e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.069266] env[61986]: DEBUG oslo_concurrency.lockutils [req-a15f3c47-c8c6-4ec2-adbd-0e9bfdc41694 req-83be40a7-8da0-4a9a-a8ef-45d3849d910f service nova] Lock "aa12919e-3891-4da9-a280-4155137864e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.069432] env[61986]: DEBUG nova.compute.manager [req-a15f3c47-c8c6-4ec2-adbd-0e9bfdc41694 req-83be40a7-8da0-4a9a-a8ef-45d3849d910f service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] No waiting events found dispatching network-vif-plugged-038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 782.069593] env[61986]: WARNING nova.compute.manager [req-a15f3c47-c8c6-4ec2-adbd-0e9bfdc41694 req-83be40a7-8da0-4a9a-a8ef-45d3849d910f service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Received unexpected event network-vif-plugged-038ca95f-90ee-4bd7-afb9-b514378d344e for instance with vm_state building and task_state spawning. [ 782.118861] env[61986]: DEBUG nova.compute.utils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.123308] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 782.123484] env[61986]: DEBUG nova.network.neutron [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 782.166857] env[61986]: DEBUG nova.policy [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4675c2084b6463bb47c93865f5cec46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f746f57f466f480c958d8b15bd2ce8fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 782.314993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.314993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.315284] env[61986]: DEBUG nova.network.neutron [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.473222] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b002f8-bee1-43fb-baee-274f06323a5e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.481801] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e12fc6-a11a-41d1-925c-3a3d32ea871a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.516731] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889a2948-a0ec-418f-8bf5-4e61e2a3c7cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.524841] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9218e9c2-0131-4f50-8f20-270c04532322 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.538344] env[61986]: DEBUG nova.compute.provider_tree [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.624946] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 782.635977] env[61986]: DEBUG nova.network.neutron [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Successfully created port: d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.884591] env[61986]: DEBUG nova.network.neutron [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.934777] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.934956] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.963112] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "076c8e91-af70-489f-89d5-f598166fafc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.963355] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "076c8e91-af70-489f-89d5-f598166fafc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.006542] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.006797] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.042608] env[61986]: DEBUG nova.scheduler.client.report [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.260558] env[61986]: DEBUG nova.network.neutron [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updating instance_info_cache with network_info: [{"id": "038ca95f-90ee-4bd7-afb9-b514378d344e", "address": "fa:16:3e:dc:61:92", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038ca95f-90", "ovs_interfaceid": "038ca95f-90ee-4bd7-afb9-b514378d344e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.552068] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.552068] env[61986]: ERROR nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Traceback (most recent call last): [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self.driver.spawn(context, instance, image_meta, [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.552068] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] vm_ref = self.build_virtual_machine(instance, [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] for vif in network_info: [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return self._sync_wrapper(fn, *args, **kwargs) [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self.wait() [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self[:] = self._gt.wait() [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return self._exit_event.wait() [ 783.552662] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] result = hub.switch() [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return self.greenlet.switch() [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] result = function(*args, **kwargs) [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] return func(*args, **kwargs) [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] raise e [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] nwinfo = self.network_api.allocate_for_instance( [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 783.553020] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] created_port_ids = self._update_ports_for_instance( [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] with excutils.save_and_reraise_exception(): [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] self.force_reraise() [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] raise self.value [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] updated_port = self._update_port( [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] _ensure_no_port_binding_failure(port) [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.553373] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] raise exception.PortBindingFailed(port_id=port['id']) [ 783.553694] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] nova.exception.PortBindingFailed: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. [ 783.553694] env[61986]: ERROR nova.compute.manager [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] [ 783.553694] env[61986]: DEBUG nova.compute.utils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.553694] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.365s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.557022] env[61986]: INFO nova.compute.claims [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.558453] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Build of instance 8bc12f2b-a632-4ffe-9d55-1a838e94931b was re-scheduled: Binding failed for port 9eb658ee-3092-4b43-af55-3e7718b0b616, please check neutron logs for more information. {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.559834] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Unplugging VIFs for instance {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.560244] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.560555] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.560861] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 783.636264] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 783.664584] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.664808] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.664968] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.665186] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.665336] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.665477] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.665678] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.665833] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.666008] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.666529] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.666672] env[61986]: DEBUG nova.virt.hardware [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.667710] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bf44ef-6f43-4dc5-bc04-43593cdd2e14 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.677088] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea718f1-df1e-439e-85d6-db744a7d4af9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.763409] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.763755] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance network_info: |[{"id": "038ca95f-90ee-4bd7-afb9-b514378d344e", "address": "fa:16:3e:dc:61:92", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038ca95f-90", "ovs_interfaceid": "038ca95f-90ee-4bd7-afb9-b514378d344e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 783.764182] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:61:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '038ca95f-90ee-4bd7-afb9-b514378d344e', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.771686] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating folder: Project (f746f57f466f480c958d8b15bd2ce8fd). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.772166] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88399f50-7cdf-46d5-85b8-c9d6e8c23311 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.784261] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created folder: Project (f746f57f466f480c958d8b15bd2ce8fd) in parent group-v252271. [ 783.784443] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating folder: Instances. Parent ref: group-v252283. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.784916] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e1f04be-ab7b-48e7-9adf-cde1b7a9f7e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.793633] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created folder: Instances in parent group-v252283. [ 783.793858] env[61986]: DEBUG oslo.service.loopingcall [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.794042] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 783.794243] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb228a8f-2f62-4299-b467-753097a5745a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.815113] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.815113] env[61986]: value = "task-1159546" [ 783.815113] env[61986]: _type = "Task" [ 783.815113] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.823138] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159546, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.094542] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.102043] env[61986]: DEBUG nova.compute.manager [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Received event network-changed-038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.102241] env[61986]: DEBUG nova.compute.manager [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Refreshing instance network info cache due to event network-changed-038ca95f-90ee-4bd7-afb9-b514378d344e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.102447] env[61986]: DEBUG oslo_concurrency.lockutils [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] Acquiring lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.102586] env[61986]: DEBUG oslo_concurrency.lockutils [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] Acquired lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.102894] env[61986]: DEBUG nova.network.neutron [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Refreshing network info cache for port 038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.265035] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.331020] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159546, 'name': CreateVM_Task, 'duration_secs': 0.363457} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.331020] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 784.331195] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.331245] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.331569] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.331821] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-753dee1e-c0c4-4353-884c-192b5127b300 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.336714] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 784.336714] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526c77bb-28e4-0e59-742e-3b67379be0a8" [ 784.336714] env[61986]: _type = "Task" [ 784.336714] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.345158] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526c77bb-28e4-0e59-742e-3b67379be0a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.457209] env[61986]: DEBUG nova.network.neutron [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Successfully updated port: d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.768027] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "refresh_cache-8bc12f2b-a632-4ffe-9d55-1a838e94931b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.768322] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61986) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.769372] env[61986]: DEBUG nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.769582] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 784.796512] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.850189] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526c77bb-28e4-0e59-742e-3b67379be0a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009803} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.850393] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.850617] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 784.850853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.850993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.851180] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.853921] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0786835-4611-4c87-b139-9018ef59df04 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.862609] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.864533] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 784.864533] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86cd5307-a575-4604-b48d-b33baf373786 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.874950] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 784.874950] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da3952-e38d-2d9f-a54e-aad6ff8bfc4c" [ 784.874950] env[61986]: _type = "Task" [ 784.874950] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.885752] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da3952-e38d-2d9f-a54e-aad6ff8bfc4c, 'name': SearchDatastore_Task, 'duration_secs': 0.009061} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.886476] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3fad2bc-a023-4b10-b5b0-449a0b3f4229 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.894955] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 784.894955] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da7f61-346b-98cb-1760-6c4aa184bf84" [ 784.894955] env[61986]: _type = "Task" [ 784.894955] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.903221] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da7f61-346b-98cb-1760-6c4aa184bf84, 'name': SearchDatastore_Task, 'duration_secs': 0.008828} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.903459] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.903703] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.903942] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1567ded-d794-4878-bb2b-7d8b6877dac9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.911687] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 784.911687] env[61986]: value = "task-1159547" [ 784.911687] env[61986]: _type = "Task" [ 784.911687] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.918036] env[61986]: DEBUG nova.network.neutron [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updated VIF entry in instance network info cache for port 038ca95f-90ee-4bd7-afb9-b514378d344e. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 784.918388] env[61986]: DEBUG nova.network.neutron [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updating instance_info_cache with network_info: [{"id": "038ca95f-90ee-4bd7-afb9-b514378d344e", "address": "fa:16:3e:dc:61:92", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038ca95f-90", "ovs_interfaceid": "038ca95f-90ee-4bd7-afb9-b514378d344e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.924440] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159547, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.961994] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.962812] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.963030] env[61986]: DEBUG nova.network.neutron [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.019967] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2befe2fc-a9c0-4381-9404-9c9f73209434 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.027549] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7ad695-d860-4bd2-bc7c-958dd70bac33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.059722] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda9c776-beb9-4f1e-a522-e6ed7d077c6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.072732] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07b58cc-a5eb-406d-b4f4-5f68321062ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.088136] env[61986]: DEBUG nova.compute.provider_tree [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.300922] env[61986]: DEBUG nova.network.neutron [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.423910] env[61986]: DEBUG oslo_concurrency.lockutils [req-b6ab86b5-8ff7-4317-b69a-2bcd987f59fa req-9ed7652c-8067-4b41-be7a-d2d67203036e service nova] Releasing lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.424358] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159547, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473991} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.424655] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.424802] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.425049] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5182306e-73a7-493a-a960-541fc3bd38ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.431459] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 785.431459] env[61986]: value = "task-1159548" [ 785.431459] env[61986]: _type = "Task" [ 785.431459] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.439272] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159548, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.513270] env[61986]: DEBUG nova.network.neutron [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.591605] env[61986]: DEBUG nova.scheduler.client.report [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.663782] env[61986]: DEBUG nova.network.neutron [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updating instance_info_cache with network_info: [{"id": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "address": "fa:16:3e:6a:56:8b", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a2fed5-6f", "ovs_interfaceid": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.803801] env[61986]: INFO nova.compute.manager [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 8bc12f2b-a632-4ffe-9d55-1a838e94931b] Took 1.03 seconds to deallocate network for instance. [ 785.945841] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159548, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060674} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.946511] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.948369] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52418b49-4584-446e-9245-031298f80e5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.976879] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.977678] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b63ab66-f111-40a1-8ecb-fb66cb68aaf9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.999279] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 785.999279] env[61986]: value = "task-1159549" [ 785.999279] env[61986]: _type = "Task" [ 785.999279] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.008140] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.100080] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.100080] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.101692] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.026s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.103434] env[61986]: INFO nova.compute.claims [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.155246] env[61986]: DEBUG nova.compute.manager [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Received event network-vif-plugged-d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.155497] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] Acquiring lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.156906] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.156906] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.156906] env[61986]: DEBUG nova.compute.manager [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] No waiting events found dispatching network-vif-plugged-d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 786.156906] env[61986]: WARNING nova.compute.manager [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Received unexpected event network-vif-plugged-d9a2fed5-6f97-43b2-9e7a-5c859c341ebe for instance with vm_state building and task_state spawning. [ 786.156906] env[61986]: DEBUG nova.compute.manager [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Received event network-changed-d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.157482] env[61986]: DEBUG nova.compute.manager [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Refreshing instance network info cache due to event network-changed-d9a2fed5-6f97-43b2-9e7a-5c859c341ebe. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 786.157482] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] Acquiring lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.166562] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.166868] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Instance network_info: |[{"id": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "address": "fa:16:3e:6a:56:8b", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a2fed5-6f", "ovs_interfaceid": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 786.167401] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] Acquired lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.167616] env[61986]: DEBUG nova.network.neutron [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Refreshing network info cache for port d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 786.168758] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:56:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9a2fed5-6f97-43b2-9e7a-5c859c341ebe', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.176023] env[61986]: DEBUG oslo.service.loopingcall [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.179066] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 786.179779] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7c6f499-655e-42cf-92bc-525b5d589ab4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.199166] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.199166] env[61986]: value = "task-1159550" [ 786.199166] env[61986]: _type = "Task" [ 786.199166] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.206378] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159550, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.511505] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159549, 'name': ReconfigVM_Task, 'duration_secs': 0.471196} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.511837] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Reconfigured VM instance instance-00000026 to attach disk [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.512544] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-08422029-2a32-44be-a7c6-c4950a73be20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.519488] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 786.519488] env[61986]: value = "task-1159551" [ 786.519488] env[61986]: _type = "Task" [ 786.519488] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.528149] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159551, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.548156] env[61986]: DEBUG nova.network.neutron [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updated VIF entry in instance network info cache for port d9a2fed5-6f97-43b2-9e7a-5c859c341ebe. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 786.548544] env[61986]: DEBUG nova.network.neutron [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updating instance_info_cache with network_info: [{"id": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "address": "fa:16:3e:6a:56:8b", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a2fed5-6f", "ovs_interfaceid": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.608622] env[61986]: DEBUG nova.compute.utils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.612603] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.612772] env[61986]: DEBUG nova.network.neutron [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 786.682771] env[61986]: DEBUG nova.policy [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90eaabe22eb740d3a16831acca706ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2e3f4d630ce4fb2a783adc56941d1ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 786.708725] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159550, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.839431] env[61986]: INFO nova.scheduler.client.report [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted allocations for instance 8bc12f2b-a632-4ffe-9d55-1a838e94931b [ 787.030941] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159551, 'name': Rename_Task, 'duration_secs': 0.141604} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.031246] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 787.031491] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65fa2718-2cfd-49ab-a7db-8bcfa2cae509 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.038603] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 787.038603] env[61986]: value = "task-1159552" [ 787.038603] env[61986]: _type = "Task" [ 787.038603] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.048779] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.055539] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a5789e9-f535-4c47-9355-47992307267c req-df01a549-4192-4edb-b662-85fd3168417f service nova] Releasing lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.118402] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.213971] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159550, 'name': CreateVM_Task, 'duration_secs': 0.611635} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.214157] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 787.214870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.215456] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.215456] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.215852] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83891270-e455-4091-9db2-21f9e05fdacf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.223650] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 787.223650] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52df2234-25b7-b34f-c5c2-210d5716c9c5" [ 787.223650] env[61986]: _type = "Task" [ 787.223650] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.235469] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52df2234-25b7-b34f-c5c2-210d5716c9c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009688} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.235752] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.235968] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.236196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.236333] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.236501] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.236937] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-607aeda0-dee7-4e5f-8e1a-7f0806c4c71e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.247302] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.247482] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 787.248245] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e2ff659-641f-4416-a551-c7448a69b756 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.256124] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 787.256124] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522de64c-4461-4cb2-24e3-febbf4639508" [ 787.256124] env[61986]: _type = "Task" [ 787.256124] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.264933] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522de64c-4461-4cb2-24e3-febbf4639508, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.296996] env[61986]: DEBUG nova.network.neutron [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Successfully created port: 56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.347055] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27191439-590a-4c60-b2da-ebfd38b0185e tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "8bc12f2b-a632-4ffe-9d55-1a838e94931b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.287s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.556975] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159552, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.579553] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e849a0-3fe9-4683-9f70-5b16739ceeb1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.589237] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42caa00b-4cf5-433a-aba6-939fe039c580 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.630640] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8772dc6-6102-4d74-b146-cb58149d09c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.640993] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2027f8-cf0f-42a4-807e-40ca26ee04b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.653407] env[61986]: DEBUG nova.compute.provider_tree [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.771143] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522de64c-4461-4cb2-24e3-febbf4639508, 'name': SearchDatastore_Task, 'duration_secs': 0.00892} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.773851] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e8d7991-c67a-44e5-b627-7c05aff1ce5d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.780281] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 787.780281] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528356d3-a893-5ddc-c604-a2b95b16fc82" [ 787.780281] env[61986]: _type = "Task" [ 787.780281] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.788899] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528356d3-a893-5ddc-c604-a2b95b16fc82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.850417] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.052678] env[61986]: DEBUG oslo_vmware.api [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159552, 'name': PowerOnVM_Task, 'duration_secs': 0.990486} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.052957] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.057088] env[61986]: INFO nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Took 6.89 seconds to spawn the instance on the hypervisor. [ 788.057088] env[61986]: DEBUG nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.057088] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3044787e-6701-400e-9063-e2d169fa3eb7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.135915] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.159120] env[61986]: DEBUG nova.scheduler.client.report [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.171318] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.175019] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.175019] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.175019] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.175019] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.175019] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.175019] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.175386] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.175386] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.175386] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.175386] env[61986]: DEBUG nova.virt.hardware [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.175386] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "2377b482-52e7-4054-8d5b-8f00587c3371" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.175568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "2377b482-52e7-4054-8d5b-8f00587c3371" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.175568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "2377b482-52e7-4054-8d5b-8f00587c3371-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.175568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "2377b482-52e7-4054-8d5b-8f00587c3371-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.175568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "2377b482-52e7-4054-8d5b-8f00587c3371-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.177389] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5566bc-f255-4fdf-802e-78dc11cfdd32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.182246] env[61986]: INFO nova.compute.manager [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Terminating instance [ 788.184925] env[61986]: DEBUG nova.compute.manager [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.185269] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 788.186218] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f1ab50-4ab2-4380-bf6d-9c1af727d626 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.195602] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47649c97-0ec0-4b18-b6f7-dd09169f03bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.206059] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 788.206059] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afcc6dee-4ca6-4bb6-ab3a-b744b409fc00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.220352] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 788.220352] env[61986]: value = "task-1159553" [ 788.220352] env[61986]: _type = "Task" [ 788.220352] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.228906] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.291809] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528356d3-a893-5ddc-c604-a2b95b16fc82, 'name': SearchDatastore_Task, 'duration_secs': 0.0175} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.292369] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.292574] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d/aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.292834] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6111e514-4a42-4c41-adce-cc4dc8bae9c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.301219] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 788.301219] env[61986]: value = "task-1159554" [ 788.301219] env[61986]: _type = "Task" [ 788.301219] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.309590] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.376429] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.576015] env[61986]: INFO nova.compute.manager [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Took 25.68 seconds to build instance. [ 788.664792] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.665513] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.670547] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.135s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.672337] env[61986]: INFO nova.compute.claims [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.732348] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159553, 'name': PowerOffVM_Task, 'duration_secs': 0.188521} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.732707] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 788.734835] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 788.734835] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d86f8efa-b506-4e10-9306-696ab33d4c7e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.801568] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 788.801721] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 788.801915] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleting the datastore file [datastore2] 2377b482-52e7-4054-8d5b-8f00587c3371 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.802315] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-966c7b4a-4ac1-42b8-b472-832dde508e6e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.820051] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 788.820051] env[61986]: value = "task-1159556" [ 788.820051] env[61986]: _type = "Task" [ 788.820051] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.820051] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159554, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.830808] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.970909] env[61986]: DEBUG nova.compute.manager [req-f2f03ab3-5c2b-468e-ab36-5d58e9137f35 req-2b7a492f-723a-44d5-b990-bab54dc9d565 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Received event network-vif-plugged-56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.971218] env[61986]: DEBUG oslo_concurrency.lockutils [req-f2f03ab3-5c2b-468e-ab36-5d58e9137f35 req-2b7a492f-723a-44d5-b990-bab54dc9d565 service nova] Acquiring lock "6c62a1e9-d60c-4720-a554-951bea4b2e18-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.971347] env[61986]: DEBUG oslo_concurrency.lockutils [req-f2f03ab3-5c2b-468e-ab36-5d58e9137f35 req-2b7a492f-723a-44d5-b990-bab54dc9d565 service nova] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.971580] env[61986]: DEBUG oslo_concurrency.lockutils [req-f2f03ab3-5c2b-468e-ab36-5d58e9137f35 req-2b7a492f-723a-44d5-b990-bab54dc9d565 service nova] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.971700] env[61986]: DEBUG nova.compute.manager [req-f2f03ab3-5c2b-468e-ab36-5d58e9137f35 req-2b7a492f-723a-44d5-b990-bab54dc9d565 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] No waiting events found dispatching network-vif-plugged-56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.971895] env[61986]: WARNING nova.compute.manager [req-f2f03ab3-5c2b-468e-ab36-5d58e9137f35 req-2b7a492f-723a-44d5-b990-bab54dc9d565 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Received unexpected event network-vif-plugged-56c4f9aa-e51e-4317-a30e-677e23707425 for instance with vm_state building and task_state spawning. [ 789.062447] env[61986]: DEBUG nova.network.neutron [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Successfully updated port: 56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 789.078190] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8e313be-a2b2-4c23-a390-b49fc1e60563 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa12919e-3891-4da9-a280-4155137864e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.571s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.177256] env[61986]: DEBUG nova.compute.utils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.184950] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.186041] env[61986]: DEBUG nova.network.neutron [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.253333] env[61986]: DEBUG nova.policy [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3ca3baaab63400684218e32dc8b8bdf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ddd77efba5d94caf81e14827bbae6ed8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 789.314369] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159554, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513686} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.314884] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d/aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.314884] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.315670] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ce5a362-8045-4636-905a-0adb3043f3f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.324380] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 789.324380] env[61986]: value = "task-1159557" [ 789.324380] env[61986]: _type = "Task" [ 789.324380] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.332343] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.337531] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.550842] env[61986]: DEBUG nova.network.neutron [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Successfully created port: dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.568293] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "refresh_cache-6c62a1e9-d60c-4720-a554-951bea4b2e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.568471] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquired lock "refresh_cache-6c62a1e9-d60c-4720-a554-951bea4b2e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.568620] env[61986]: DEBUG nova.network.neutron [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.585449] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.685754] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.840120] env[61986]: DEBUG oslo_vmware.api [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.550979} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.840120] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.840120] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 789.840120] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 789.840120] env[61986]: INFO nova.compute.manager [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Took 1.65 seconds to destroy the instance on the hypervisor. [ 789.840517] env[61986]: DEBUG oslo.service.loopingcall [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.841908] env[61986]: DEBUG nova.compute.manager [-] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.842031] env[61986]: DEBUG nova.network.neutron [-] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 789.846017] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079319} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.846017] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.846017] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcd1dbd-c879-4f33-9fe5-318e7649d306 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.875389] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d/aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.877479] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8c09ba4-a5cb-4a29-8ad6-824df5516bf9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.899461] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 789.899461] env[61986]: value = "task-1159558" [ 789.899461] env[61986]: _type = "Task" [ 789.899461] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.909901] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159558, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.105407] env[61986]: DEBUG nova.network.neutron [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 790.107700] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.113969] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa05848-b513-4670-b52a-ec829bcebff7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.121827] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b51a2a8-be14-4079-ac21-8f5b1cde8752 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.160422] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdad6df9-5218-4f1d-89d3-19865663dcb2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.168385] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e905c41f-011d-4883-b70a-936f32a88599 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.184844] env[61986]: DEBUG nova.compute.provider_tree [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.287785] env[61986]: DEBUG nova.network.neutron [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Updating instance_info_cache with network_info: [{"id": "56c4f9aa-e51e-4317-a30e-677e23707425", "address": "fa:16:3e:eb:4b:4a", "network": {"id": "538c773c-a7c4-4e52-8006-b9a237236aa3", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-77316531-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e3f4d630ce4fb2a783adc56941d1ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "309d7cfa-b4da-4eec-9f4b-2e10d215fac7", "external-id": "nsx-vlan-transportzone-285", "segmentation_id": 285, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56c4f9aa-e5", "ovs_interfaceid": "56c4f9aa-e51e-4317-a30e-677e23707425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.359059] env[61986]: DEBUG nova.compute.manager [req-f23a66c9-577f-4f27-86bb-bc721e0e1684 req-d7d3bc0a-c054-4e60-b1b0-5aa79c88241a service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Received event network-vif-deleted-91d1d769-094d-4ae2-84a2-868652bd2b82 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.359293] env[61986]: INFO nova.compute.manager [req-f23a66c9-577f-4f27-86bb-bc721e0e1684 req-d7d3bc0a-c054-4e60-b1b0-5aa79c88241a service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Neutron deleted interface 91d1d769-094d-4ae2-84a2-868652bd2b82; detaching it from the instance and deleting it from the info cache [ 790.359466] env[61986]: DEBUG nova.network.neutron [req-f23a66c9-577f-4f27-86bb-bc721e0e1684 req-d7d3bc0a-c054-4e60-b1b0-5aa79c88241a service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.410032] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159558, 'name': ReconfigVM_Task, 'duration_secs': 0.289153} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.410032] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Reconfigured VM instance instance-00000027 to attach disk [datastore2] aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d/aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.410268] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8eb0d2b4-1418-4cc9-ab21-6476cf93f541 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.417547] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 790.417547] env[61986]: value = "task-1159563" [ 790.417547] env[61986]: _type = "Task" [ 790.417547] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.424879] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159563, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.690342] env[61986]: DEBUG nova.scheduler.client.report [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.695865] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.698613] env[61986]: DEBUG nova.network.neutron [-] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.731155] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:40:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='754ce345-2135-4db7-a907-0c2bf294269e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2075393374',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.731326] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.731534] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.731786] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.731972] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.732532] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.732532] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.732659] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.732811] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.733069] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.733287] env[61986]: DEBUG nova.virt.hardware [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.734565] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfec633-4cf7-4f2b-a195-0eb4602526bd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.742868] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66168056-676d-42b6-865d-afeb30b686d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.788771] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Releasing lock "refresh_cache-6c62a1e9-d60c-4720-a554-951bea4b2e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.789101] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Instance network_info: |[{"id": "56c4f9aa-e51e-4317-a30e-677e23707425", "address": "fa:16:3e:eb:4b:4a", "network": {"id": "538c773c-a7c4-4e52-8006-b9a237236aa3", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-77316531-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e3f4d630ce4fb2a783adc56941d1ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "309d7cfa-b4da-4eec-9f4b-2e10d215fac7", "external-id": "nsx-vlan-transportzone-285", "segmentation_id": 285, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56c4f9aa-e5", "ovs_interfaceid": "56c4f9aa-e51e-4317-a30e-677e23707425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 790.789533] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:4b:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '309d7cfa-b4da-4eec-9f4b-2e10d215fac7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56c4f9aa-e51e-4317-a30e-677e23707425', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.796971] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Creating folder: Project (c2e3f4d630ce4fb2a783adc56941d1ce). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 790.797629] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b93f6526-28ca-4f9d-8b14-0d508f3fa437 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.807960] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Created folder: Project (c2e3f4d630ce4fb2a783adc56941d1ce) in parent group-v252271. [ 790.808245] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Creating folder: Instances. Parent ref: group-v252290. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 790.808430] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c34bcf1-29cf-4c00-9e98-ffde8d8f9276 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.817661] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Created folder: Instances in parent group-v252290. [ 790.817804] env[61986]: DEBUG oslo.service.loopingcall [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.817994] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 790.818218] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb656732-2b31-4a88-b829-24164ba316b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.837722] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.837722] env[61986]: value = "task-1159566" [ 790.837722] env[61986]: _type = "Task" [ 790.837722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.845482] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159566, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.862580] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6446c044-a5b5-4c3f-8ac4-7cfe61888dac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.872117] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c3f0df-3c2f-4754-8f3d-07c00185ebdc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.895961] env[61986]: DEBUG nova.compute.manager [req-f23a66c9-577f-4f27-86bb-bc721e0e1684 req-d7d3bc0a-c054-4e60-b1b0-5aa79c88241a service nova] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Detach interface failed, port_id=91d1d769-094d-4ae2-84a2-868652bd2b82, reason: Instance 2377b482-52e7-4054-8d5b-8f00587c3371 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 790.928507] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159563, 'name': Rename_Task, 'duration_secs': 0.143832} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.928837] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 790.929096] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-197c9331-a187-4d51-b0c2-09f2c84afb69 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.935583] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 790.935583] env[61986]: value = "task-1159567" [ 790.935583] env[61986]: _type = "Task" [ 790.935583] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.944791] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.012199] env[61986]: DEBUG nova.compute.manager [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Received event network-changed-56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 791.012199] env[61986]: DEBUG nova.compute.manager [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Refreshing instance network info cache due to event network-changed-56c4f9aa-e51e-4317-a30e-677e23707425. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 791.013136] env[61986]: DEBUG oslo_concurrency.lockutils [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] Acquiring lock "refresh_cache-6c62a1e9-d60c-4720-a554-951bea4b2e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.013461] env[61986]: DEBUG oslo_concurrency.lockutils [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] Acquired lock "refresh_cache-6c62a1e9-d60c-4720-a554-951bea4b2e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.013733] env[61986]: DEBUG nova.network.neutron [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Refreshing network info cache for port 56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 791.109814] env[61986]: DEBUG nova.network.neutron [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Successfully updated port: dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.201894] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.202438] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.205017] env[61986]: INFO nova.compute.manager [-] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Took 1.36 seconds to deallocate network for instance. [ 791.205501] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.653s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.206968] env[61986]: INFO nova.compute.claims [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.349311] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159566, 'name': CreateVM_Task, 'duration_secs': 0.390199} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.349569] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 791.350316] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.350479] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.350787] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.351053] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fad23e1-5e2a-4e37-945a-6939f6bf872d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.356048] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 791.356048] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f5caf1-f7b8-5e87-bc08-ed55b7965eff" [ 791.356048] env[61986]: _type = "Task" [ 791.356048] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.363640] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f5caf1-f7b8-5e87-bc08-ed55b7965eff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.446035] env[61986]: DEBUG oslo_vmware.api [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159567, 'name': PowerOnVM_Task, 'duration_secs': 0.508328} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.446417] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 791.446639] env[61986]: INFO nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Took 7.81 seconds to spawn the instance on the hypervisor. [ 791.446838] env[61986]: DEBUG nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.447666] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5676d3a6-7bde-47ca-8cd3-d41fc0b93b0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.613069] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.613306] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.613680] env[61986]: DEBUG nova.network.neutron [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.707279] env[61986]: DEBUG nova.compute.utils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.708696] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.712211] env[61986]: DEBUG nova.network.neutron [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 791.722377] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.725824] env[61986]: DEBUG nova.network.neutron [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Updated VIF entry in instance network info cache for port 56c4f9aa-e51e-4317-a30e-677e23707425. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.725824] env[61986]: DEBUG nova.network.neutron [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Updating instance_info_cache with network_info: [{"id": "56c4f9aa-e51e-4317-a30e-677e23707425", "address": "fa:16:3e:eb:4b:4a", "network": {"id": "538c773c-a7c4-4e52-8006-b9a237236aa3", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-77316531-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2e3f4d630ce4fb2a783adc56941d1ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "309d7cfa-b4da-4eec-9f4b-2e10d215fac7", "external-id": "nsx-vlan-transportzone-285", "segmentation_id": 285, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56c4f9aa-e5", "ovs_interfaceid": "56c4f9aa-e51e-4317-a30e-677e23707425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.771695] env[61986]: DEBUG nova.policy [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94bb4fe585bb42339a01194fa2a68bac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '034de7dfeeb145c0a1e90b111c15e36f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 791.865416] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f5caf1-f7b8-5e87-bc08-ed55b7965eff, 'name': SearchDatastore_Task, 'duration_secs': 0.020528} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.868740] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.868955] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.869272] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.869434] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.869653] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.870260] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbfaf970-8d5f-407e-8374-210d9c137b6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.880757] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.881043] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.881675] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc143325-d7c3-47c8-9313-7bf86df57cb3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.886600] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 791.886600] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d40fc8-b736-1cec-a43c-2ad27a467899" [ 791.886600] env[61986]: _type = "Task" [ 791.886600] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.893907] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d40fc8-b736-1cec-a43c-2ad27a467899, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.967087] env[61986]: INFO nova.compute.manager [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Took 27.16 seconds to build instance. [ 792.094989] env[61986]: DEBUG nova.network.neutron [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Successfully created port: a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.160153] env[61986]: DEBUG nova.network.neutron [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.217359] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.229881] env[61986]: DEBUG oslo_concurrency.lockutils [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] Releasing lock "refresh_cache-6c62a1e9-d60c-4720-a554-951bea4b2e18" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.229881] env[61986]: DEBUG nova.compute.manager [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Received event network-vif-plugged-dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.229881] env[61986]: DEBUG oslo_concurrency.lockutils [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] Acquiring lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.229881] env[61986]: DEBUG oslo_concurrency.lockutils [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.229881] env[61986]: DEBUG oslo_concurrency.lockutils [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.230162] env[61986]: DEBUG nova.compute.manager [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] No waiting events found dispatching network-vif-plugged-dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 792.230162] env[61986]: WARNING nova.compute.manager [req-99facc6d-0ae6-4ebb-b1f1-af2ea3c772ba req-a5a9e30e-967d-4e59-b93e-83c9edf6fe64 service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Received unexpected event network-vif-plugged-dd125b45-dffb-4a25-af78-4a9326408be0 for instance with vm_state building and task_state spawning. [ 792.382669] env[61986]: DEBUG nova.network.neutron [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.402020] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d40fc8-b736-1cec-a43c-2ad27a467899, 'name': SearchDatastore_Task, 'duration_secs': 0.013993} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.402020] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89c45285-6d3c-4282-9c89-d9bad0ea8297 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.407536] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 792.407536] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52404f1f-4527-44f8-4173-d510fd68dbd7" [ 792.407536] env[61986]: _type = "Task" [ 792.407536] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.422019] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52404f1f-4527-44f8-4173-d510fd68dbd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.470217] env[61986]: DEBUG oslo_concurrency.lockutils [None req-54a021d7-d988-473c-aad0-82800dbd3987 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.563s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.596165] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059f8fad-ad69-4706-9365-b6a1bd2073e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.604237] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bffe305-62eb-44c3-8e1e-b3576f77a0ec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.635542] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ef9d10-524d-43b1-9176-010a56498e63 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.644781] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9038b2-8e80-44a5-a748-45374578e555 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.660128] env[61986]: DEBUG nova.compute.provider_tree [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.885226] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.885578] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Instance network_info: |[{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.885999] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:7e:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd125b45-dffb-4a25-af78-4a9326408be0', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.893636] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Creating folder: Project (ddd77efba5d94caf81e14827bbae6ed8). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.893850] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a1ea7c5-1484-4177-842e-8ee27c07bc09 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.905500] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Created folder: Project (ddd77efba5d94caf81e14827bbae6ed8) in parent group-v252271. [ 792.905684] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Creating folder: Instances. Parent ref: group-v252293. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.905903] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8793a282-f856-4f17-ab1f-88b47243e731 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.915439] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Created folder: Instances in parent group-v252293. [ 792.915576] env[61986]: DEBUG oslo.service.loopingcall [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.916311] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.916311] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-900cb973-b3ad-44ac-abef-83ab14aa5716 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.935179] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52404f1f-4527-44f8-4173-d510fd68dbd7, 'name': SearchDatastore_Task, 'duration_secs': 0.012695} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.935800] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.936065] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6c62a1e9-d60c-4720-a554-951bea4b2e18/6c62a1e9-d60c-4720-a554-951bea4b2e18.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 792.936369] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c8b1c3c-9bee-4f46-b166-f095f4cfcb1b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.939515] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.939515] env[61986]: value = "task-1159570" [ 792.939515] env[61986]: _type = "Task" [ 792.939515] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.943559] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 792.943559] env[61986]: value = "task-1159571" [ 792.943559] env[61986]: _type = "Task" [ 792.943559] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.949481] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159570, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.954092] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.973025] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.041244] env[61986]: DEBUG nova.compute.manager [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Received event network-changed-dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 793.041413] env[61986]: DEBUG nova.compute.manager [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Refreshing instance network info cache due to event network-changed-dd125b45-dffb-4a25-af78-4a9326408be0. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 793.041649] env[61986]: DEBUG oslo_concurrency.lockutils [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] Acquiring lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.041863] env[61986]: DEBUG oslo_concurrency.lockutils [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] Acquired lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.042148] env[61986]: DEBUG nova.network.neutron [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Refreshing network info cache for port dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.051897] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.053312] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.163376] env[61986]: DEBUG nova.scheduler.client.report [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.231651] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.256648] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.257020] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.257205] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.257686] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.257686] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.257849] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.258159] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.258349] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.258550] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.258749] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.258962] env[61986]: DEBUG nova.virt.hardware [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.260042] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102958ef-0c11-4785-bea1-6280e19936cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.270130] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d9eb83-fd6a-4ea5-b829-3a090f697877 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.454264] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159570, 'name': CreateVM_Task, 'duration_secs': 0.476652} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.457657] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.457994] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159571, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.458776] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.458965] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.459331] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.459662] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95764360-8beb-4c1d-80ac-d3b8b3d771e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.464636] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 793.464636] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522760b7-5963-f700-5119-048acf4b14ea" [ 793.464636] env[61986]: _type = "Task" [ 793.464636] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.474911] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522760b7-5963-f700-5119-048acf4b14ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.495629] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.668490] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.669036] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.671865] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.108s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.673232] env[61986]: INFO nova.compute.claims [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.879974] env[61986]: DEBUG nova.network.neutron [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updated VIF entry in instance network info cache for port dd125b45-dffb-4a25-af78-4a9326408be0. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.880377] env[61986]: DEBUG nova.network.neutron [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.935530] env[61986]: DEBUG nova.network.neutron [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Successfully updated port: a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.956700] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676793} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.957164] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6c62a1e9-d60c-4720-a554-951bea4b2e18/6c62a1e9-d60c-4720-a554-951bea4b2e18.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 793.957298] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.957563] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d44e4e91-29d6-40bd-aeea-a8df70b5e3f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.967312] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 793.967312] env[61986]: value = "task-1159573" [ 793.967312] env[61986]: _type = "Task" [ 793.967312] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.980426] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522760b7-5963-f700-5119-048acf4b14ea, 'name': SearchDatastore_Task, 'duration_secs': 0.025259} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.983590] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.984201] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.984201] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.984201] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.984432] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.985039] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159573, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.985269] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d47d9853-6986-4291-89c7-669857b536ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.000736] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.001041] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 794.003567] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daaacebb-ab74-41cb-9774-0dc05372f19b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.009875] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 794.009875] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f5f451-7ea5-677a-a494-e9eb9ddad5c1" [ 794.009875] env[61986]: _type = "Task" [ 794.009875] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.017763] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f5f451-7ea5-677a-a494-e9eb9ddad5c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.181046] env[61986]: DEBUG nova.compute.utils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.182579] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.182743] env[61986]: DEBUG nova.network.neutron [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.228794] env[61986]: DEBUG nova.policy [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 794.383296] env[61986]: DEBUG oslo_concurrency.lockutils [req-e44a60d6-54e5-4889-b89a-0f16910e46a2 req-45c28c92-15fa-4d13-95f0-44e18696562c service nova] Releasing lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.442186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "refresh_cache-6e0697d4-d862-497c-9bdd-dd1ef2d4272b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.442345] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired lock "refresh_cache-6e0697d4-d862-497c-9bdd-dd1ef2d4272b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.442501] env[61986]: DEBUG nova.network.neutron [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 794.481194] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159573, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.171916} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.481497] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.482550] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dec91ed-30fe-4f8c-b000-bdd11869739b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.514709] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 6c62a1e9-d60c-4720-a554-951bea4b2e18/6c62a1e9-d60c-4720-a554-951bea4b2e18.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.514709] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdb3c354-4a8e-43de-8527-3e78cd0ccbcb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.530553] env[61986]: DEBUG nova.network.neutron [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Successfully created port: 4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.542974] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f5f451-7ea5-677a-a494-e9eb9ddad5c1, 'name': SearchDatastore_Task, 'duration_secs': 0.022486} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.545082] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 794.545082] env[61986]: value = "task-1159574" [ 794.545082] env[61986]: _type = "Task" [ 794.545082] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.545082] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dad45733-f496-4aef-8dd0-106312875e5a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.552071] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 794.552071] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520e05a5-4dc1-d2ae-33fd-d57f76229b76" [ 794.552071] env[61986]: _type = "Task" [ 794.552071] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.555706] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159574, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.562962] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520e05a5-4dc1-d2ae-33fd-d57f76229b76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.685603] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.976897] env[61986]: DEBUG nova.network.neutron [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 795.034417] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdf7c21-d363-48b1-b5cb-da9ed3c0b303 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.042182] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd499d2-0f72-486a-8f4c-207743e8617f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.080792] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48f5115-24bd-4a87-94a5-5d76cec3c6d4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.086973] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.096405] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520e05a5-4dc1-d2ae-33fd-d57f76229b76, 'name': SearchDatastore_Task, 'duration_secs': 0.021847} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.097120] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.097406] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 795.097681] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4772c75d-c5f2-4780-aa05-30cdc58cd2ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.103119] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87051932-cc4f-40e7-933f-9adac0f4064c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.108542] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 795.108542] env[61986]: value = "task-1159575" [ 795.108542] env[61986]: _type = "Task" [ 795.108542] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.119413] env[61986]: DEBUG nova.compute.provider_tree [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.126810] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.168340] env[61986]: DEBUG nova.network.neutron [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Updating instance_info_cache with network_info: [{"id": "a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3", "address": "fa:16:3e:f2:0f:b7", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.20", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9fa57e7-ab", "ovs_interfaceid": "a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.299427] env[61986]: DEBUG nova.compute.manager [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Received event network-vif-plugged-a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 795.299427] env[61986]: DEBUG oslo_concurrency.lockutils [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] Acquiring lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.299427] env[61986]: DEBUG oslo_concurrency.lockutils [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.299427] env[61986]: DEBUG oslo_concurrency.lockutils [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.299427] env[61986]: DEBUG nova.compute.manager [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] No waiting events found dispatching network-vif-plugged-a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 795.299744] env[61986]: WARNING nova.compute.manager [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Received unexpected event network-vif-plugged-a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 for instance with vm_state building and task_state spawning. [ 795.299744] env[61986]: DEBUG nova.compute.manager [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Received event network-changed-a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 795.299810] env[61986]: DEBUG nova.compute.manager [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Refreshing instance network info cache due to event network-changed-a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 795.300017] env[61986]: DEBUG oslo_concurrency.lockutils [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] Acquiring lock "refresh_cache-6e0697d4-d862-497c-9bdd-dd1ef2d4272b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.557891] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159574, 'name': ReconfigVM_Task, 'duration_secs': 0.968629} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.558815] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 6c62a1e9-d60c-4720-a554-951bea4b2e18/6c62a1e9-d60c-4720-a554-951bea4b2e18.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.559763] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2115fe6-cec3-45fe-ac3d-2d8d45f3e426 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.567879] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 795.567879] env[61986]: value = "task-1159577" [ 795.567879] env[61986]: _type = "Task" [ 795.567879] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.577676] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159577, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.620418] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159575, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.627778] env[61986]: DEBUG nova.scheduler.client.report [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.670722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Releasing lock "refresh_cache-6e0697d4-d862-497c-9bdd-dd1ef2d4272b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.671153] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Instance network_info: |[{"id": "a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3", "address": "fa:16:3e:f2:0f:b7", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.20", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9fa57e7-ab", "ovs_interfaceid": "a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 795.671557] env[61986]: DEBUG oslo_concurrency.lockutils [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] Acquired lock "refresh_cache-6e0697d4-d862-497c-9bdd-dd1ef2d4272b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.671797] env[61986]: DEBUG nova.network.neutron [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Refreshing network info cache for port a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.673226] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:0f:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.682288] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Creating folder: Project (034de7dfeeb145c0a1e90b111c15e36f). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 795.683341] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d9ff262-ebd6-40b2-8f5c-4208a178d535 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.695723] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.698553] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Created folder: Project (034de7dfeeb145c0a1e90b111c15e36f) in parent group-v252271. [ 795.698772] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Creating folder: Instances. Parent ref: group-v252296. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 795.699285] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95158d22-fb2a-4ef4-ba70-4150d06cafd7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.709785] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Created folder: Instances in parent group-v252296. [ 795.710466] env[61986]: DEBUG oslo.service.loopingcall [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.710466] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 795.710647] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7261232a-71ce-499a-866e-a7dbd950722c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.734288] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.734288] env[61986]: value = "task-1159580" [ 795.734288] env[61986]: _type = "Task" [ 795.734288] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.740454] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.740764] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.740943] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.741145] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.741465] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.741569] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.741818] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.741982] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.742184] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.742433] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.742511] env[61986]: DEBUG nova.virt.hardware [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.744065] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44947660-a925-47c3-9f86-6793b1645ac2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.752432] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159580, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.756414] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4283b3e1-e067-4d61-b642-b6392176d53c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.026108] env[61986]: DEBUG nova.network.neutron [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Successfully updated port: 4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.080024] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159577, 'name': Rename_Task, 'duration_secs': 0.21377} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.080024] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 796.080024] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbfb1d1a-15a4-4906-aa48-a0f0def622af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.086409] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 796.086409] env[61986]: value = "task-1159581" [ 796.086409] env[61986]: _type = "Task" [ 796.086409] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.096676] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159581, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.119969] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683059} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.120119] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.120367] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.120619] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29cc5b0e-de46-4636-9e19-df54ac15210d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.129935] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 796.129935] env[61986]: value = "task-1159582" [ 796.129935] env[61986]: _type = "Task" [ 796.129935] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.133836] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.134631] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.137236] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.793s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.138730] env[61986]: INFO nova.compute.claims [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.146568] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159582, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.244522] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159580, 'name': CreateVM_Task, 'duration_secs': 0.488753} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.244699] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 796.245464] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.245625] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.245934] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.246201] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f23afb3a-9c88-48c3-82de-52c88c335f19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.250959] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 796.250959] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e2fcbf-b87e-5e39-7f98-807c6c41ad9a" [ 796.250959] env[61986]: _type = "Task" [ 796.250959] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.261546] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e2fcbf-b87e-5e39-7f98-807c6c41ad9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.407298] env[61986]: DEBUG nova.network.neutron [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Updated VIF entry in instance network info cache for port a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.407647] env[61986]: DEBUG nova.network.neutron [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Updating instance_info_cache with network_info: [{"id": "a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3", "address": "fa:16:3e:f2:0f:b7", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.20", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9fa57e7-ab", "ovs_interfaceid": "a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.528527] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.528682] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.528833] env[61986]: DEBUG nova.network.neutron [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.596640] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159581, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.640752] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159582, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.643149] env[61986]: DEBUG nova.compute.utils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.649600] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.649600] env[61986]: DEBUG nova.network.neutron [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.710189] env[61986]: DEBUG nova.policy [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'acf18a4c613a46e093c2685de6357e3e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32de5b569f744582b9f1d0dcb4cd7c1d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 796.761924] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e2fcbf-b87e-5e39-7f98-807c6c41ad9a, 'name': SearchDatastore_Task, 'duration_secs': 0.00939} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.762346] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.762609] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.762857] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.763015] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.763198] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.763490] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18972998-5b83-45fe-9a5f-b1fa67ef2747 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.772630] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.772863] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 796.773710] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9ac608a-c149-4ec7-be61-d6ddab313524 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.779305] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 796.779305] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f7c9f-a4e1-56e8-e0ef-321e15bbc897" [ 796.779305] env[61986]: _type = "Task" [ 796.779305] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.787162] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f7c9f-a4e1-56e8-e0ef-321e15bbc897, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.910490] env[61986]: DEBUG oslo_concurrency.lockutils [req-fd72a2d9-4f74-44dd-9fc4-aa3e5125e116 req-f61a161f-7e82-41c6-9355-913d455af598 service nova] Releasing lock "refresh_cache-6e0697d4-d862-497c-9bdd-dd1ef2d4272b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.063690] env[61986]: DEBUG nova.network.neutron [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Successfully created port: 8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.079918] env[61986]: DEBUG nova.network.neutron [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.096688] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159581, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.142222] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159582, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.896293} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.142557] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.143417] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a631fbf6-ba50-4ff7-b6b2-ae6806e2d1bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.146672] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.172771] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.175875] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2f5435b-7eef-4e61-a7ca-2b98e390543a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.207483] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 797.207483] env[61986]: value = "task-1159583" [ 797.207483] env[61986]: _type = "Task" [ 797.207483] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.221019] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159583, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.274985] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.275947] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.294079] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f7c9f-a4e1-56e8-e0ef-321e15bbc897, 'name': SearchDatastore_Task, 'duration_secs': 0.014636} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.295624] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8a4abaf-f473-4a4f-82fa-dc79e5d92261 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.302629] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 797.302629] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522552c6-a606-2959-dbdf-46aa28546b33" [ 797.302629] env[61986]: _type = "Task" [ 797.302629] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.314976] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522552c6-a606-2959-dbdf-46aa28546b33, 'name': SearchDatastore_Task, 'duration_secs': 0.009597} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.315907] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.315907] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 6e0697d4-d862-497c-9bdd-dd1ef2d4272b/6e0697d4-d862-497c-9bdd-dd1ef2d4272b.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 797.315907] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-564c5389-fff2-41c2-a61d-d209f2018f1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.328519] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 797.328519] env[61986]: value = "task-1159584" [ 797.328519] env[61986]: _type = "Task" [ 797.328519] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.339518] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159584, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.492739] env[61986]: DEBUG nova.compute.manager [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-vif-plugged-4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.492991] env[61986]: DEBUG oslo_concurrency.lockutils [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.493210] env[61986]: DEBUG oslo_concurrency.lockutils [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.493388] env[61986]: DEBUG oslo_concurrency.lockutils [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.493570] env[61986]: DEBUG nova.compute.manager [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] No waiting events found dispatching network-vif-plugged-4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 797.493735] env[61986]: WARNING nova.compute.manager [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received unexpected event network-vif-plugged-4c8749d8-2c4f-455b-9e9f-c4dc823cc295 for instance with vm_state building and task_state spawning. [ 797.493908] env[61986]: DEBUG nova.compute.manager [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-changed-4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.494072] env[61986]: DEBUG nova.compute.manager [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing instance network info cache due to event network-changed-4c8749d8-2c4f-455b-9e9f-c4dc823cc295. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 797.494260] env[61986]: DEBUG oslo_concurrency.lockutils [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.515136] env[61986]: DEBUG nova.network.neutron [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.604282] env[61986]: DEBUG oslo_vmware.api [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159581, 'name': PowerOnVM_Task, 'duration_secs': 1.192301} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.604626] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 797.604867] env[61986]: INFO nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Took 9.47 seconds to spawn the instance on the hypervisor. [ 797.605181] env[61986]: DEBUG nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.606201] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6bd682-c205-4e89-ba17-37e5b0bf5153 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.673778] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d1b745-68d5-4411-bba5-1b02eb3ac017 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.695771] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca00b33-f173-4647-b9db-2b406fa6d24e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.752980] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d01877a-4b73-4a48-855c-e85f543ac2a1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.770656] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9f0d46-aa38-405c-ba59-ba6bd660ca1b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.778115] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159583, 'name': ReconfigVM_Task, 'duration_secs': 0.408599} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.778624] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.785017] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d7b78f8-49db-4ff7-90d3-0da3f2dcf15f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.801325] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.801579] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 797.801754] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Rebuilding the list of instances to heal {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 797.805402] env[61986]: DEBUG nova.compute.provider_tree [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.814555] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 797.814555] env[61986]: value = "task-1159586" [ 797.814555] env[61986]: _type = "Task" [ 797.814555] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.831263] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159586, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.844702] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159584, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.019903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.020845] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Instance network_info: |[{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.020845] env[61986]: DEBUG oslo_concurrency.lockutils [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.021088] env[61986]: DEBUG nova.network.neutron [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing network info cache for port 4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.022117] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:b8:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e0cfc48-d93b-4477-8082-69a2f7aa7701', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c8749d8-2c4f-455b-9e9f-c4dc823cc295', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.029975] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Creating folder: Project (617d3b641cdc4d1b815b5edf7e970d62). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.033253] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ff0d916-ee06-4440-876d-59894334481c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.046235] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Created folder: Project (617d3b641cdc4d1b815b5edf7e970d62) in parent group-v252271. [ 798.046441] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Creating folder: Instances. Parent ref: group-v252299. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.046689] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1b26a37-f21f-4c75-871e-62e4cd1f7fbc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.055903] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Created folder: Instances in parent group-v252299. [ 798.056173] env[61986]: DEBUG oslo.service.loopingcall [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.056367] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.056577] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45148fa9-5378-42ef-a718-28b502aa8a56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.078044] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.078044] env[61986]: value = "task-1159589" [ 798.078044] env[61986]: _type = "Task" [ 798.078044] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.085957] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159589, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.139473] env[61986]: INFO nova.compute.manager [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Took 28.98 seconds to build instance. [ 798.187439] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.215600] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.215914] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.216105] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.216303] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.216501] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.216580] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.216783] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.216939] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.217353] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.217560] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.217743] env[61986]: DEBUG nova.virt.hardware [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.218734] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0681f24e-0048-457d-be8e-f6207c9bdcfc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.226950] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23a0e02-9e37-45dd-880f-5f9984054dac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.296886] env[61986]: DEBUG nova.network.neutron [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updated VIF entry in instance network info cache for port 4c8749d8-2c4f-455b-9e9f-c4dc823cc295. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.297386] env[61986]: DEBUG nova.network.neutron [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.306722] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 798.306889] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 798.307041] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 798.307179] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 798.307275] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 798.311122] env[61986]: DEBUG nova.scheduler.client.report [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.326289] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159586, 'name': Rename_Task, 'duration_secs': 0.501067} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.327548] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.327548] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5806708b-3c22-4074-b84d-126f07e5732b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.336125] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 798.336125] env[61986]: value = "task-1159590" [ 798.336125] env[61986]: _type = "Task" [ 798.336125] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.350706] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159584, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545206} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.350706] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 6e0697d4-d862-497c-9bdd-dd1ef2d4272b/6e0697d4-d862-497c-9bdd-dd1ef2d4272b.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 798.350706] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.350706] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.350706] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-849a3d06-b104-4f8c-82af-b35d17488afe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.352537] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.352685] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.352836] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 798.353017] env[61986]: DEBUG nova.objects.instance [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lazy-loading 'info_cache' on Instance uuid 2377b482-52e7-4054-8d5b-8f00587c3371 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.359619] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 798.359619] env[61986]: value = "task-1159591" [ 798.359619] env[61986]: _type = "Task" [ 798.359619] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.368516] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159591, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.588212] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159589, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.642554] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d26b920-0fd6-4d97-b9bc-8defd79e92bf tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.064s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.801331] env[61986]: DEBUG oslo_concurrency.lockutils [req-97497378-adc3-48d7-ad87-44431b89bd59 req-4aa435bf-d9b2-4392-96b1-c7620d67748c service nova] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.816276] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.817049] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.819808] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.333s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.821560] env[61986]: INFO nova.compute.claims [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.849631] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159590, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.870580] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159591, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107538} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.874017] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.874017] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca162c47-287a-47a9-a9d5-78a46c691ed4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.899822] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 6e0697d4-d862-497c-9bdd-dd1ef2d4272b/6e0697d4-d862-497c-9bdd-dd1ef2d4272b.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.901567] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86f681b3-abe3-448d-84e0-ea9f26ada58b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.923077] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 798.923077] env[61986]: value = "task-1159592" [ 798.923077] env[61986]: _type = "Task" [ 798.923077] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.934589] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159592, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.953605] env[61986]: DEBUG nova.network.neutron [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Successfully updated port: 8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.089721] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159589, 'name': CreateVM_Task, 'duration_secs': 0.623256} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.089798] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 799.090565] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.090740] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.091099] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.091301] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-007ef34e-e693-4064-a6cd-395bcc6e6129 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.096132] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 799.096132] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52303c1a-e8f5-87d3-5884-7643234fcef1" [ 799.096132] env[61986]: _type = "Task" [ 799.096132] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.104170] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52303c1a-e8f5-87d3-5884-7643234fcef1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.144888] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.327079] env[61986]: DEBUG nova.compute.utils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.331367] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.331367] env[61986]: DEBUG nova.network.neutron [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 799.355034] env[61986]: DEBUG oslo_vmware.api [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159590, 'name': PowerOnVM_Task, 'duration_secs': 0.895481} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.355034] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.355034] env[61986]: INFO nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Took 8.66 seconds to spawn the instance on the hypervisor. [ 799.355034] env[61986]: DEBUG nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.357103] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9956fa3a-20b3-46c6-87f0-890f3a14e205 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.395545] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.404459] env[61986]: DEBUG nova.policy [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e34c464681464270a5ae8c02b6e3fb75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ea5fc77300041f5aeccb8773300ef0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 799.432890] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159592, 'name': ReconfigVM_Task, 'duration_secs': 0.457594} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.434158] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 6e0697d4-d862-497c-9bdd-dd1ef2d4272b/6e0697d4-d862-497c-9bdd-dd1ef2d4272b.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.434158] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78626640-1f21-4430-9569-8499bb56c1c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.440548] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 799.440548] env[61986]: value = "task-1159593" [ 799.440548] env[61986]: _type = "Task" [ 799.440548] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.450085] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159593, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.457039] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "refresh_cache-4bb588b3-c5e1-4619-b77b-6734d6516170" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.457039] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquired lock "refresh_cache-4bb588b3-c5e1-4619-b77b-6734d6516170" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.457258] env[61986]: DEBUG nova.network.neutron [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.561614] env[61986]: DEBUG nova.compute.manager [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Received event network-vif-plugged-8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.562425] env[61986]: DEBUG oslo_concurrency.lockutils [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] Acquiring lock "4bb588b3-c5e1-4619-b77b-6734d6516170-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.563334] env[61986]: DEBUG oslo_concurrency.lockutils [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.565602] env[61986]: DEBUG oslo_concurrency.lockutils [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.565602] env[61986]: DEBUG nova.compute.manager [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] No waiting events found dispatching network-vif-plugged-8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.565602] env[61986]: WARNING nova.compute.manager [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Received unexpected event network-vif-plugged-8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 for instance with vm_state building and task_state spawning. [ 799.565602] env[61986]: DEBUG nova.compute.manager [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Received event network-changed-8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.565602] env[61986]: DEBUG nova.compute.manager [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Refreshing instance network info cache due to event network-changed-8fa7deb8-c0a2-45a5-98f9-1c07972e34f6. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 799.565845] env[61986]: DEBUG oslo_concurrency.lockutils [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] Acquiring lock "refresh_cache-4bb588b3-c5e1-4619-b77b-6734d6516170" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.612811] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52303c1a-e8f5-87d3-5884-7643234fcef1, 'name': SearchDatastore_Task, 'duration_secs': 0.012733} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.613158] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.613392] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.613632] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.613778] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.613950] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.614217] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86e8cf79-e920-4e76-a684-d7ef58576c71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.623478] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.623722] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.627440] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37e25e39-3c4c-485d-b143-f857e39efeb6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.633411] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 799.633411] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528722f7-ea2e-e464-ba30-6883cb1b880e" [ 799.633411] env[61986]: _type = "Task" [ 799.633411] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.641528] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528722f7-ea2e-e464-ba30-6883cb1b880e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.678465] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.704956] env[61986]: DEBUG nova.network.neutron [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Successfully created port: 0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.832022] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.884043] env[61986]: INFO nova.compute.manager [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Took 28.83 seconds to build instance. [ 799.952887] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159593, 'name': Rename_Task, 'duration_secs': 0.139167} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.957228] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 799.957808] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73ea2cab-1aaa-4578-812f-7e42c9ac7dba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.967449] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 799.967449] env[61986]: value = "task-1159595" [ 799.967449] env[61986]: _type = "Task" [ 799.967449] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.985680] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159595, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.000235] env[61986]: DEBUG nova.network.neutron [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.058389] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.148192] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528722f7-ea2e-e464-ba30-6883cb1b880e, 'name': SearchDatastore_Task, 'duration_secs': 0.012753} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.149078] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66ccc0ec-474c-42d2-9356-ade701e751aa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.156966] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 800.156966] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d30c6a-18f1-7436-b4fb-4899ed6d90ce" [ 800.156966] env[61986]: _type = "Task" [ 800.156966] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.166366] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d30c6a-18f1-7436-b4fb-4899ed6d90ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.260934] env[61986]: DEBUG nova.network.neutron [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Updating instance_info_cache with network_info: [{"id": "8fa7deb8-c0a2-45a5-98f9-1c07972e34f6", "address": "fa:16:3e:02:c9:2d", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.189", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa7deb8-c0", "ovs_interfaceid": "8fa7deb8-c0a2-45a5-98f9-1c07972e34f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.281592] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d13b1d-d461-4539-aa96-76da07a9a8e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.292622] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e3a66d-6fe7-4573-9d9c-5f547300cdc2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.326327] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805a7f2f-96f3-46e9-96c3-fb9611336241 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.334096] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf931a3-aedd-4ae0-825a-72516adad425 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.354753] env[61986]: DEBUG nova.compute.provider_tree [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.387479] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1ed6cf7e-92af-4059-98a4-706c223b8895 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.654s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.482456] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159595, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.563255] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-2377b482-52e7-4054-8d5b-8f00587c3371" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.563344] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 800.563564] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.563720] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.563862] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.564012] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.564164] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.564301] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.564419] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 800.564554] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.667098] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d30c6a-18f1-7436-b4fb-4899ed6d90ce, 'name': SearchDatastore_Task, 'duration_secs': 0.011262} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.667478] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.667653] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 977799b4-2793-4513-9447-483146fc7ac4/977799b4-2793-4513-9447-483146fc7ac4.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 800.667921] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7254b2b-4733-493d-863a-a1e46c1a69db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.674890] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 800.674890] env[61986]: value = "task-1159596" [ 800.674890] env[61986]: _type = "Task" [ 800.674890] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.682823] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.768726] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Releasing lock "refresh_cache-4bb588b3-c5e1-4619-b77b-6734d6516170" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.769085] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Instance network_info: |[{"id": "8fa7deb8-c0a2-45a5-98f9-1c07972e34f6", "address": "fa:16:3e:02:c9:2d", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.189", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa7deb8-c0", "ovs_interfaceid": "8fa7deb8-c0a2-45a5-98f9-1c07972e34f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.769406] env[61986]: DEBUG oslo_concurrency.lockutils [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] Acquired lock "refresh_cache-4bb588b3-c5e1-4619-b77b-6734d6516170" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.769564] env[61986]: DEBUG nova.network.neutron [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Refreshing network info cache for port 8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.770739] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:c9:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fa7deb8-c0a2-45a5-98f9-1c07972e34f6', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.778779] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Creating folder: Project (32de5b569f744582b9f1d0dcb4cd7c1d). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.780780] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be84e075-95f7-4456-a203-8a1a9bb8f6a5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.792150] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Created folder: Project (32de5b569f744582b9f1d0dcb4cd7c1d) in parent group-v252271. [ 800.793033] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Creating folder: Instances. Parent ref: group-v252303. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.793033] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ff55119-cf0a-4273-a573-e937687a2354 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.801837] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Created folder: Instances in parent group-v252303. [ 800.802081] env[61986]: DEBUG oslo.service.loopingcall [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.802263] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.802458] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-058b662b-fb04-4166-85d7-abdeb7dc1563 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.822577] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.822577] env[61986]: value = "task-1159599" [ 800.822577] env[61986]: _type = "Task" [ 800.822577] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.830225] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159599, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.845067] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.858710] env[61986]: DEBUG nova.scheduler.client.report [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.874087] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.874087] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.874087] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.874479] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.874875] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.875300] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.875749] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.877223] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.877223] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.877223] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.877223] env[61986]: DEBUG nova.virt.hardware [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.878026] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219a1847-a969-4a25-8e01-f246b07c556c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.886861] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77138bd-a5db-4f61-8452-7b263f6e8c55 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.891055] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.978235] env[61986]: DEBUG oslo_vmware.api [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159595, 'name': PowerOnVM_Task, 'duration_secs': 0.537124} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.978992] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 800.978992] env[61986]: INFO nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Took 7.75 seconds to spawn the instance on the hypervisor. [ 800.978992] env[61986]: DEBUG nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 800.980222] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b2998f-11d2-4f1b-bf53-da384498809b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.067283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.185254] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159596, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.335416] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159599, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.349755] env[61986]: DEBUG nova.network.neutron [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Successfully updated port: 0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.364984] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.365705] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.369380] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.993s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.372999] env[61986]: INFO nova.compute.claims [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.412765] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "6c62a1e9-d60c-4720-a554-951bea4b2e18" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.412765] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.412765] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "6c62a1e9-d60c-4720-a554-951bea4b2e18-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.412765] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.413053] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.416066] env[61986]: INFO nova.compute.manager [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Terminating instance [ 801.418518] env[61986]: DEBUG nova.compute.manager [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 801.418818] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 801.419747] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af2f4a5-c2ff-4515-845a-dd638188daeb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.428392] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.430742] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 801.430988] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24323392-73a2-4109-8c23-128d6f858419 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.439248] env[61986]: DEBUG oslo_vmware.api [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 801.439248] env[61986]: value = "task-1159600" [ 801.439248] env[61986]: _type = "Task" [ 801.439248] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.450944] env[61986]: DEBUG oslo_vmware.api [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.499703] env[61986]: INFO nova.compute.manager [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Took 25.98 seconds to build instance. [ 801.555054] env[61986]: DEBUG nova.compute.manager [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Stashing vm_state: active {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 801.690411] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.964807} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.690855] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 977799b4-2793-4513-9447-483146fc7ac4/977799b4-2793-4513-9447-483146fc7ac4.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.691841] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.691841] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f44b56e-edf9-43e2-ad47-f476fa75c13e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.701206] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 801.701206] env[61986]: value = "task-1159602" [ 801.701206] env[61986]: _type = "Task" [ 801.701206] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.713877] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159602, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.730288] env[61986]: DEBUG nova.network.neutron [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Updated VIF entry in instance network info cache for port 8fa7deb8-c0a2-45a5-98f9-1c07972e34f6. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.730288] env[61986]: DEBUG nova.network.neutron [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Updating instance_info_cache with network_info: [{"id": "8fa7deb8-c0a2-45a5-98f9-1c07972e34f6", "address": "fa:16:3e:02:c9:2d", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.189", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fa7deb8-c0", "ovs_interfaceid": "8fa7deb8-c0a2-45a5-98f9-1c07972e34f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.817641] env[61986]: DEBUG nova.compute.manager [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Received event network-vif-plugged-0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 801.817757] env[61986]: DEBUG oslo_concurrency.lockutils [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] Acquiring lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.818199] env[61986]: DEBUG oslo_concurrency.lockutils [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.818199] env[61986]: DEBUG oslo_concurrency.lockutils [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.818308] env[61986]: DEBUG nova.compute.manager [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] No waiting events found dispatching network-vif-plugged-0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 801.818576] env[61986]: WARNING nova.compute.manager [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Received unexpected event network-vif-plugged-0495bb14-032d-43ea-9462-1d1d0d326723 for instance with vm_state building and task_state spawning. [ 801.818722] env[61986]: DEBUG nova.compute.manager [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Received event network-changed-0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 801.820992] env[61986]: DEBUG nova.compute.manager [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Refreshing instance network info cache due to event network-changed-0495bb14-032d-43ea-9462-1d1d0d326723. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 801.820992] env[61986]: DEBUG oslo_concurrency.lockutils [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] Acquiring lock "refresh_cache-c1a6ffac-99bd-492c-99e5-f7c46b352d8e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.820992] env[61986]: DEBUG oslo_concurrency.lockutils [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] Acquired lock "refresh_cache-c1a6ffac-99bd-492c-99e5-f7c46b352d8e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.820992] env[61986]: DEBUG nova.network.neutron [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Refreshing network info cache for port 0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.836051] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159599, 'name': CreateVM_Task, 'duration_secs': 0.663811} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.836235] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 801.836985] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.837197] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.837572] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.838111] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-184c2499-c896-46c9-8007-060115443b96 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.844093] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 801.844093] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f731e5-c91e-6f08-2103-98b48f857a00" [ 801.844093] env[61986]: _type = "Task" [ 801.844093] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.853834] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "refresh_cache-c1a6ffac-99bd-492c-99e5-f7c46b352d8e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.854138] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f731e5-c91e-6f08-2103-98b48f857a00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.880665] env[61986]: DEBUG nova.compute.utils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.882830] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.883015] env[61986]: DEBUG nova.network.neutron [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 801.955901] env[61986]: DEBUG oslo_vmware.api [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159600, 'name': PowerOffVM_Task, 'duration_secs': 0.229463} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.957711] env[61986]: DEBUG nova.policy [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd541ca28165c499285d0683ee82ea3a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5b5abb4c676745be8d77ad83e7e0a9ab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 801.960411] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 801.960784] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 801.961193] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32d82d21-7943-43c6-b429-c893208a5572 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.003455] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0a3b454-aaae-4aab-884f-13e0eade6ab1 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.043s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.042215] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 802.042215] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 802.042215] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Deleting the datastore file [datastore2] 6c62a1e9-d60c-4720-a554-951bea4b2e18 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 802.042215] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0158c376-8510-4b5b-962b-61483092d2d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.049563] env[61986]: DEBUG oslo_vmware.api [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for the task: (returnval){ [ 802.049563] env[61986]: value = "task-1159604" [ 802.049563] env[61986]: _type = "Task" [ 802.049563] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.058728] env[61986]: DEBUG oslo_vmware.api [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.076255] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.212472] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159602, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071742} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.212968] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.214122] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344989ce-e8e6-46db-bbbd-1e18674bb409 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.239554] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 977799b4-2793-4513-9447-483146fc7ac4/977799b4-2793-4513-9447-483146fc7ac4.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.240499] env[61986]: DEBUG oslo_concurrency.lockutils [req-1c28605e-1f21-4a60-8420-71dc9671c716 req-1b90f213-f43d-4820-a8c1-ed5855aaf21e service nova] Releasing lock "refresh_cache-4bb588b3-c5e1-4619-b77b-6734d6516170" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.240984] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-effb62ed-7b74-4e8e-a0a0-83ed15628688 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.267428] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 802.267428] env[61986]: value = "task-1159605" [ 802.267428] env[61986]: _type = "Task" [ 802.267428] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.281460] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.358173] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f731e5-c91e-6f08-2103-98b48f857a00, 'name': SearchDatastore_Task, 'duration_secs': 0.016713} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.358173] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.358173] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.358173] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.358404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.358404] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.358497] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b264575c-b7a5-4c8b-aadf-29a2ecad6dd2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.369604] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.369903] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 802.371197] env[61986]: DEBUG nova.network.neutron [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.372952] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96fa7403-0369-47b2-b468-564dc3ecbfc0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.379735] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 802.379735] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aeecd0-3095-2378-ef25-7c8a8f7c8683" [ 802.379735] env[61986]: _type = "Task" [ 802.379735] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.389961] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.400434] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aeecd0-3095-2378-ef25-7c8a8f7c8683, 'name': SearchDatastore_Task, 'duration_secs': 0.014885} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.404681] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbefa064-9708-4932-95a0-87002f7373f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.407816] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 802.407816] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52af0dab-b6d5-ee25-8136-15905f6649eb" [ 802.407816] env[61986]: _type = "Task" [ 802.407816] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.421452] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52af0dab-b6d5-ee25-8136-15905f6649eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.438915] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "4f13b629-e2a7-4668-9d77-eb638078e246" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.439576] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "4f13b629-e2a7-4668-9d77-eb638078e246" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.493604] env[61986]: DEBUG nova.network.neutron [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.504769] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.548834] env[61986]: DEBUG nova.network.neutron [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Successfully created port: 9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.564435] env[61986]: DEBUG oslo_vmware.api [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Task: {'id': task-1159604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.402314} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.564695] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.564869] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 802.565295] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 802.565373] env[61986]: INFO nova.compute.manager [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Took 1.15 seconds to destroy the instance on the hypervisor. [ 802.565594] env[61986]: DEBUG oslo.service.loopingcall [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.565776] env[61986]: DEBUG nova.compute.manager [-] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.565916] env[61986]: DEBUG nova.network.neutron [-] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.779140] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159605, 'name': ReconfigVM_Task, 'duration_secs': 0.310132} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.782080] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 977799b4-2793-4513-9447-483146fc7ac4/977799b4-2793-4513-9447-483146fc7ac4.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.782910] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f16f76ee-33fa-4c71-983c-73dd790d0bb5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.790739] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 802.790739] env[61986]: value = "task-1159606" [ 802.790739] env[61986]: _type = "Task" [ 802.790739] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.800293] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159606, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.846935] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23196e8-e67f-479e-9639-2c6f7d902cea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.855711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7614f9b-1330-4e3a-b732-8aee8ca624e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.891868] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b09c440-e599-409c-bb93-d19366e25bd8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.904452] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c01c04-337f-4b08-9386-b041e06b8d6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.918970] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52af0dab-b6d5-ee25-8136-15905f6649eb, 'name': SearchDatastore_Task, 'duration_secs': 0.014306} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.928412] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.928733] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 4bb588b3-c5e1-4619-b77b-6734d6516170/4bb588b3-c5e1-4619-b77b-6734d6516170.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 802.931014] env[61986]: DEBUG nova.compute.provider_tree [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.934521] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-065605f9-52a4-43a6-b198-0ec3292360d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.939439] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 802.939439] env[61986]: value = "task-1159607" [ 802.939439] env[61986]: _type = "Task" [ 802.939439] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.949646] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.996686] env[61986]: DEBUG oslo_concurrency.lockutils [req-001ac8b5-ae36-442a-a1f6-b6fb65e7d39f req-b051d290-798e-4be0-b6d4-6f44939766c7 service nova] Releasing lock "refresh_cache-c1a6ffac-99bd-492c-99e5-f7c46b352d8e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.997215] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "refresh_cache-c1a6ffac-99bd-492c-99e5-f7c46b352d8e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.997402] env[61986]: DEBUG nova.network.neutron [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.041445] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.304030] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159606, 'name': Rename_Task, 'duration_secs': 0.145823} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.304380] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 803.304708] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bbdd554-66a3-462c-805e-e605b1b98b95 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.315808] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 803.315808] env[61986]: value = "task-1159608" [ 803.315808] env[61986]: _type = "Task" [ 803.315808] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.327571] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.335546] env[61986]: DEBUG nova.network.neutron [-] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.400571] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 803.429757] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.430031] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.430219] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.430417] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.430576] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.430722] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.430958] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.431151] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.431335] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.431512] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.431695] env[61986]: DEBUG nova.virt.hardware [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.432642] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df3c8d2-8e47-43bc-9079-fae0f40efc07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.436115] env[61986]: DEBUG nova.scheduler.client.report [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.448100] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05573b6-00c5-4e51-8a26-d17e52b8b026 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.456577] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510251} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.457314] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 4bb588b3-c5e1-4619-b77b-6734d6516170/4bb588b3-c5e1-4619-b77b-6734d6516170.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 803.457575] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.457863] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9d7290e-9a1a-4fbe-95e4-5fb8b1d4d439 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.475648] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 803.475648] env[61986]: value = "task-1159609" [ 803.475648] env[61986]: _type = "Task" [ 803.475648] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.484718] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.538306] env[61986]: DEBUG nova.network.neutron [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.687319] env[61986]: DEBUG nova.network.neutron [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Updating instance_info_cache with network_info: [{"id": "0495bb14-032d-43ea-9462-1d1d0d326723", "address": "fa:16:3e:38:1b:e1", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0495bb14-03", "ovs_interfaceid": "0495bb14-032d-43ea-9462-1d1d0d326723", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.826933] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159608, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.839641] env[61986]: INFO nova.compute.manager [-] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Took 1.27 seconds to deallocate network for instance. [ 803.852576] env[61986]: DEBUG nova.compute.manager [req-3e6717a6-2823-4e17-beff-dab01dbbdc1c req-1ee3a26c-6b2b-453d-b926-a9d81338d9e2 service nova] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Received event network-vif-deleted-56c4f9aa-e51e-4317-a30e-677e23707425 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.940940] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.941493] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.944341] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.837s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.945785] env[61986]: INFO nova.compute.claims [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.986483] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069206} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.986753] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.987599] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc1b84e-d838-4265-8b3a-f8d59d6a8c9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.011017] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 4bb588b3-c5e1-4619-b77b-6734d6516170/4bb588b3-c5e1-4619-b77b-6734d6516170.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.011598] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff0f3acf-983a-416a-a271-dba6f1152a3d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.032039] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 804.032039] env[61986]: value = "task-1159611" [ 804.032039] env[61986]: _type = "Task" [ 804.032039] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.040809] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.189870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "refresh_cache-c1a6ffac-99bd-492c-99e5-f7c46b352d8e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.190230] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Instance network_info: |[{"id": "0495bb14-032d-43ea-9462-1d1d0d326723", "address": "fa:16:3e:38:1b:e1", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0495bb14-03", "ovs_interfaceid": "0495bb14-032d-43ea-9462-1d1d0d326723", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.190667] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:1b:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0954fad3-d24d-496c-83e6-a09d3cb556fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0495bb14-032d-43ea-9462-1d1d0d326723', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.198877] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Creating folder: Project (0ea5fc77300041f5aeccb8773300ef0e). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.199099] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-80a202c7-3ec5-405c-9194-19f0e823cac1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.210983] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Created folder: Project (0ea5fc77300041f5aeccb8773300ef0e) in parent group-v252271. [ 804.211184] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Creating folder: Instances. Parent ref: group-v252306. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.211425] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2077228-3c7f-4917-942f-0b08edddda53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.221793] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Created folder: Instances in parent group-v252306. [ 804.222038] env[61986]: DEBUG oslo.service.loopingcall [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.222229] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 804.222430] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f3b65ee-6f46-435a-8aa0-dc28b5bc96d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.242893] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.242893] env[61986]: value = "task-1159614" [ 804.242893] env[61986]: _type = "Task" [ 804.242893] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.258475] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159614, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.329499] env[61986]: DEBUG oslo_vmware.api [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159608, 'name': PowerOnVM_Task, 'duration_secs': 0.525901} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.329878] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 804.330195] env[61986]: INFO nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Took 8.63 seconds to spawn the instance on the hypervisor. [ 804.331434] env[61986]: DEBUG nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.331434] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb81fed8-919f-49b3-9c1f-19a80c26e42e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.351007] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.450560] env[61986]: DEBUG nova.compute.utils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.457073] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 804.457073] env[61986]: DEBUG nova.network.neutron [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 804.496050] env[61986]: DEBUG nova.compute.manager [req-611b079e-7275-45f8-a59f-3ee5cf8afa1a req-7d4136cb-caa6-4bf5-b903-26bfd6bbfb07 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Received event network-vif-plugged-9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 804.496333] env[61986]: DEBUG oslo_concurrency.lockutils [req-611b079e-7275-45f8-a59f-3ee5cf8afa1a req-7d4136cb-caa6-4bf5-b903-26bfd6bbfb07 service nova] Acquiring lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.496507] env[61986]: DEBUG oslo_concurrency.lockutils [req-611b079e-7275-45f8-a59f-3ee5cf8afa1a req-7d4136cb-caa6-4bf5-b903-26bfd6bbfb07 service nova] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.496672] env[61986]: DEBUG oslo_concurrency.lockutils [req-611b079e-7275-45f8-a59f-3ee5cf8afa1a req-7d4136cb-caa6-4bf5-b903-26bfd6bbfb07 service nova] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.496838] env[61986]: DEBUG nova.compute.manager [req-611b079e-7275-45f8-a59f-3ee5cf8afa1a req-7d4136cb-caa6-4bf5-b903-26bfd6bbfb07 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] No waiting events found dispatching network-vif-plugged-9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.496995] env[61986]: WARNING nova.compute.manager [req-611b079e-7275-45f8-a59f-3ee5cf8afa1a req-7d4136cb-caa6-4bf5-b903-26bfd6bbfb07 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Received unexpected event network-vif-plugged-9ffa532a-af81-442f-9253-1c78d86bbd6f for instance with vm_state building and task_state spawning. [ 804.499129] env[61986]: DEBUG nova.policy [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ab2d0df5b67405eaff415ae4f714248', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1625ea8d35c4e1ea7291998932e6b17', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 804.543803] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159611, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.695941] env[61986]: DEBUG nova.network.neutron [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Successfully updated port: 9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.755918] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159614, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.756974] env[61986]: DEBUG nova.network.neutron [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Successfully created port: f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.854614] env[61986]: INFO nova.compute.manager [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Took 27.32 seconds to build instance. [ 804.959262] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.046673] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159611, 'name': ReconfigVM_Task, 'duration_secs': 0.904053} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.046959] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 4bb588b3-c5e1-4619-b77b-6734d6516170/4bb588b3-c5e1-4619-b77b-6734d6516170.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.047764] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-116bbc26-8391-4b4a-bb71-3d48df156ac4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.058161] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 805.058161] env[61986]: value = "task-1159615" [ 805.058161] env[61986]: _type = "Task" [ 805.058161] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.067574] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159615, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.199032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.199032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquired lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.199032] env[61986]: DEBUG nova.network.neutron [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.257779] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159614, 'name': CreateVM_Task, 'duration_secs': 0.770115} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.258372] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.259202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.259484] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.260075] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.260491] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e19b69e4-2ac0-4ebc-8682-391a5ac9dd5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.275048] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 805.275048] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5297047a-c870-4c95-7c65-e8ea66eb96ec" [ 805.275048] env[61986]: _type = "Task" [ 805.275048] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.286427] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5297047a-c870-4c95-7c65-e8ea66eb96ec, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.287076] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.287443] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.287805] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.288084] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.288405] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.288770] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64ee9da4-6cd0-4829-be09-386df7f44f20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.303023] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.303023] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 805.303023] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bbcddbd-5bba-4933-8a6e-25874a13a65c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.306872] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 805.306872] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523f05af-71a8-a396-30e4-f4363d9aa275" [ 805.306872] env[61986]: _type = "Task" [ 805.306872] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.318980] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523f05af-71a8-a396-30e4-f4363d9aa275, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.361136] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee3937de-3369-4bd9-9ae5-641ec2b95af2 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "977799b4-2793-4513-9447-483146fc7ac4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.052s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.456107] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d306719-cb04-4763-bd56-16db5d6a4764 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.463238] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f78cd1-2661-4e75-9b5e-78546104855c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.503352] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ceca283-ebed-49b8-9406-44e1ea1eeb7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.512728] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383d7c8a-f833-4ae5-929e-a34bdf7f84b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.527598] env[61986]: DEBUG nova.compute.provider_tree [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.568184] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159615, 'name': Rename_Task, 'duration_secs': 0.144827} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.568465] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.568706] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0497c221-3a14-498a-b79f-30f10d46ea14 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.575954] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 805.575954] env[61986]: value = "task-1159616" [ 805.575954] env[61986]: _type = "Task" [ 805.575954] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.584321] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.766079] env[61986]: DEBUG nova.network.neutron [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 805.818069] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523f05af-71a8-a396-30e4-f4363d9aa275, 'name': SearchDatastore_Task, 'duration_secs': 0.010832} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.818797] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f88b51e9-2c92-45af-ab44-ae12550f2c7d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.825189] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 805.825189] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c584e6-43b3-b3b3-da2e-ebfdd59adbcb" [ 805.825189] env[61986]: _type = "Task" [ 805.825189] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.834798] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c584e6-43b3-b3b3-da2e-ebfdd59adbcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.863760] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.004057] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.028563] env[61986]: DEBUG nova.network.neutron [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updating instance_info_cache with network_info: [{"id": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "address": "fa:16:3e:32:23:4b", "network": {"id": "c15ae3c7-4631-4c01-9027-64898fd0163c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1866800563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b5abb4c676745be8d77ad83e7e0a9ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ffa532a-af", "ovs_interfaceid": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.034572] env[61986]: DEBUG nova.scheduler.client.report [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.040906] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.041272] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.041673] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.041673] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.041777] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.041894] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.042097] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.042837] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.042837] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.042837] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.043078] env[61986]: DEBUG nova.virt.hardware [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.044468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9855f951-9739-42c0-a2c6-2c7f981e82bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.054832] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44573ca0-053b-463b-ba8b-2208a0bdc217 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.063249] env[61986]: DEBUG nova.compute.manager [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-changed-4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.063249] env[61986]: DEBUG nova.compute.manager [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing instance network info cache due to event network-changed-4c8749d8-2c4f-455b-9e9f-c4dc823cc295. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.063417] env[61986]: DEBUG oslo_concurrency.lockutils [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.063459] env[61986]: DEBUG oslo_concurrency.lockutils [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.063672] env[61986]: DEBUG nova.network.neutron [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing network info cache for port 4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.088160] env[61986]: DEBUG oslo_vmware.api [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159616, 'name': PowerOnVM_Task, 'duration_secs': 0.50949} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.089426] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.089426] env[61986]: INFO nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Took 7.90 seconds to spawn the instance on the hypervisor. [ 806.089426] env[61986]: DEBUG nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.090021] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ef5b9a-96d8-437a-aaa2-9bbecb70711e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.337641] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c584e6-43b3-b3b3-da2e-ebfdd59adbcb, 'name': SearchDatastore_Task, 'duration_secs': 0.011458} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.337641] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.337641] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] c1a6ffac-99bd-492c-99e5-f7c46b352d8e/c1a6ffac-99bd-492c-99e5-f7c46b352d8e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 806.337641] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b37e1da-6f92-420e-9d70-21733cb3e0a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.346259] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 806.346259] env[61986]: value = "task-1159618" [ 806.346259] env[61986]: _type = "Task" [ 806.346259] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.355945] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159618, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.403009] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.500352] env[61986]: DEBUG nova.network.neutron [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updated VIF entry in instance network info cache for port 4c8749d8-2c4f-455b-9e9f-c4dc823cc295. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.500723] env[61986]: DEBUG nova.network.neutron [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.534490] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Releasing lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.534898] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Instance network_info: |[{"id": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "address": "fa:16:3e:32:23:4b", "network": {"id": "c15ae3c7-4631-4c01-9027-64898fd0163c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1866800563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b5abb4c676745be8d77ad83e7e0a9ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ffa532a-af", "ovs_interfaceid": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 806.535445] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:23:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ffa532a-af81-442f-9253-1c78d86bbd6f', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.550266] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Creating folder: Project (5b5abb4c676745be8d77ad83e7e0a9ab). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 806.553587] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.555161] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.559921] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d962b770-acca-49d5-adb5-c21cb34607d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.566573] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.844s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.568538] env[61986]: DEBUG nova.objects.instance [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lazy-loading 'resources' on Instance uuid 2377b482-52e7-4054-8d5b-8f00587c3371 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 806.585664] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Created folder: Project (5b5abb4c676745be8d77ad83e7e0a9ab) in parent group-v252271. [ 806.585664] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Creating folder: Instances. Parent ref: group-v252309. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 806.585664] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5ad8060-c269-46e0-9998-c852c6cc160c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.605773] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Created folder: Instances in parent group-v252309. [ 806.605773] env[61986]: DEBUG oslo.service.loopingcall [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.606040] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 806.607950] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee81c0f0-0e06-43c6-b986-87f5f3d558cc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.625453] env[61986]: INFO nova.compute.manager [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Took 27.08 seconds to build instance. [ 806.634169] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.634169] env[61986]: value = "task-1159621" [ 806.634169] env[61986]: _type = "Task" [ 806.634169] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.646320] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159621, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.646643] env[61986]: DEBUG nova.network.neutron [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Successfully updated port: f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.696067] env[61986]: DEBUG nova.compute.manager [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Received event network-changed-9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.696392] env[61986]: DEBUG nova.compute.manager [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Refreshing instance network info cache due to event network-changed-9ffa532a-af81-442f-9253-1c78d86bbd6f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.696863] env[61986]: DEBUG oslo_concurrency.lockutils [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] Acquiring lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.697311] env[61986]: DEBUG oslo_concurrency.lockutils [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] Acquired lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.697548] env[61986]: DEBUG nova.network.neutron [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Refreshing network info cache for port 9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.859109] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159618, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.009191] env[61986]: DEBUG oslo_concurrency.lockutils [req-721108a7-8461-4434-bb13-6618bc422d5f req-6d4f5fc5-4538-4953-b4f0-9e0f46d89fee service nova] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.069435] env[61986]: DEBUG nova.compute.utils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.070760] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.070950] env[61986]: DEBUG nova.network.neutron [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 807.114418] env[61986]: DEBUG nova.policy [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6881bdd48c4f0ea6bcb7d2d3746d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e9d0c35a4de4f5f9829a7f3c88fde92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 807.141993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5cab44c9-064b-47ef-952e-ee64720c3736 tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.112s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.153031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "refresh_cache-8438d2a7-975c-4d1f-8906-f960bfadfe39" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.153031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquired lock "refresh_cache-8438d2a7-975c-4d1f-8906-f960bfadfe39" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.153031] env[61986]: DEBUG nova.network.neutron [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.154106] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159621, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.357019] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159618, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543873} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.361768] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] c1a6ffac-99bd-492c-99e5-f7c46b352d8e/c1a6ffac-99bd-492c-99e5-f7c46b352d8e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.362016] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.362528] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c7e86b7-83f2-4878-abc5-85b428e73285 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.373345] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 807.373345] env[61986]: value = "task-1159622" [ 807.373345] env[61986]: _type = "Task" [ 807.373345] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.390089] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.456853] env[61986]: DEBUG nova.network.neutron [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Successfully created port: 86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.490060] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "4bb588b3-c5e1-4619-b77b-6734d6516170" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.490499] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.490723] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "4bb588b3-c5e1-4619-b77b-6734d6516170-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.491049] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.491339] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.493709] env[61986]: INFO nova.compute.manager [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Terminating instance [ 807.498549] env[61986]: DEBUG nova.compute.manager [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.498549] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.501246] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643ad240-9694-4419-8f23-73ee85a54576 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.510618] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.512880] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5d1079d-3792-4af5-a998-b50e055c0329 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.522013] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 807.522013] env[61986]: value = "task-1159623" [ 807.522013] env[61986]: _type = "Task" [ 807.522013] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.526099] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbca194d-5999-4e8d-b762-92b2ccd919eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.539876] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044c9307-3284-4b11-94ab-0bf0d60b8023 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.543693] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159623, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.574679] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 807.581357] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbabd0d3-e0cc-4e3a-8c3f-482a474641c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.591221] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4bf044-6556-4d0e-9dec-e110025f7fd8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.608217] env[61986]: DEBUG nova.compute.provider_tree [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.642523] env[61986]: DEBUG nova.network.neutron [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updated VIF entry in instance network info cache for port 9ffa532a-af81-442f-9253-1c78d86bbd6f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 807.642892] env[61986]: DEBUG nova.network.neutron [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updating instance_info_cache with network_info: [{"id": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "address": "fa:16:3e:32:23:4b", "network": {"id": "c15ae3c7-4631-4c01-9027-64898fd0163c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1866800563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b5abb4c676745be8d77ad83e7e0a9ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ffa532a-af", "ovs_interfaceid": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.644345] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.650192] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159621, 'name': CreateVM_Task, 'duration_secs': 0.665656} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.650534] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 807.651176] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.651931] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.651931] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.652208] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6d9ff1c-4749-4832-8842-1067f91bfc1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.659041] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 807.659041] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c1c624-b80b-a23a-1adb-d3250b9a5ca1" [ 807.659041] env[61986]: _type = "Task" [ 807.659041] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.668968] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c1c624-b80b-a23a-1adb-d3250b9a5ca1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.707525] env[61986]: DEBUG nova.network.neutron [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.887285] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083397} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.887584] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.891394] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be39771-0bb8-4f7a-be80-cbe80501d01d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.915203] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] c1a6ffac-99bd-492c-99e5-f7c46b352d8e/c1a6ffac-99bd-492c-99e5-f7c46b352d8e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.915505] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ba18c93-2d11-498a-b9eb-448b351db669 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.931906] env[61986]: DEBUG nova.network.neutron [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Updating instance_info_cache with network_info: [{"id": "f1530791-8284-4c8e-a87a-44b3e6145516", "address": "fa:16:3e:e0:69:52", "network": {"id": "ddf922cf-f9d3-449b-b44c-4e89b9f8a306", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1883633649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1625ea8d35c4e1ea7291998932e6b17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1530791-82", "ovs_interfaceid": "f1530791-8284-4c8e-a87a-44b3e6145516", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.938231] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 807.938231] env[61986]: value = "task-1159624" [ 807.938231] env[61986]: _type = "Task" [ 807.938231] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.947046] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.033392] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159623, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.090799] env[61986]: DEBUG nova.compute.manager [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Received event network-vif-plugged-f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.091062] env[61986]: DEBUG oslo_concurrency.lockutils [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] Acquiring lock "8438d2a7-975c-4d1f-8906-f960bfadfe39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.091292] env[61986]: DEBUG oslo_concurrency.lockutils [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.091470] env[61986]: DEBUG oslo_concurrency.lockutils [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.091672] env[61986]: DEBUG nova.compute.manager [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] No waiting events found dispatching network-vif-plugged-f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.091895] env[61986]: WARNING nova.compute.manager [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Received unexpected event network-vif-plugged-f1530791-8284-4c8e-a87a-44b3e6145516 for instance with vm_state building and task_state spawning. [ 808.092077] env[61986]: DEBUG nova.compute.manager [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Received event network-changed-f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.092286] env[61986]: DEBUG nova.compute.manager [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Refreshing instance network info cache due to event network-changed-f1530791-8284-4c8e-a87a-44b3e6145516. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 808.092467] env[61986]: DEBUG oslo_concurrency.lockutils [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] Acquiring lock "refresh_cache-8438d2a7-975c-4d1f-8906-f960bfadfe39" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.113490] env[61986]: DEBUG nova.scheduler.client.report [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.151289] env[61986]: DEBUG oslo_concurrency.lockutils [req-a7159681-3ae3-451f-8b53-87011a5f816e req-2ae15420-f03f-4d70-bc36-7cbb7e1552f6 service nova] Releasing lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.167685] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.171759] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c1c624-b80b-a23a-1adb-d3250b9a5ca1, 'name': SearchDatastore_Task, 'duration_secs': 0.05103} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.172071] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.172317] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.172566] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.172723] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.172918] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.173214] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-468a2139-1757-43f4-aced-1dd51b8c6b5f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.182903] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.183201] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 808.184137] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cec9144-9b6f-498b-a1fc-af30c9f7617d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.189961] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 808.189961] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e29d25-73f3-a938-f492-146d4b99829b" [ 808.189961] env[61986]: _type = "Task" [ 808.189961] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.199636] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e29d25-73f3-a938-f492-146d4b99829b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.435267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Releasing lock "refresh_cache-8438d2a7-975c-4d1f-8906-f960bfadfe39" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.435729] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Instance network_info: |[{"id": "f1530791-8284-4c8e-a87a-44b3e6145516", "address": "fa:16:3e:e0:69:52", "network": {"id": "ddf922cf-f9d3-449b-b44c-4e89b9f8a306", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1883633649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1625ea8d35c4e1ea7291998932e6b17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1530791-82", "ovs_interfaceid": "f1530791-8284-4c8e-a87a-44b3e6145516", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.436112] env[61986]: DEBUG oslo_concurrency.lockutils [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] Acquired lock "refresh_cache-8438d2a7-975c-4d1f-8906-f960bfadfe39" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.436319] env[61986]: DEBUG nova.network.neutron [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Refreshing network info cache for port f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.437673] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:69:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f87a752-ebb0-49a4-a67b-e356fa45b89b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1530791-8284-4c8e-a87a-44b3e6145516', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.445901] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Creating folder: Project (a1625ea8d35c4e1ea7291998932e6b17). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.446536] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ba96059-6adc-4d39-ad75-ac97fdf9cfbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.460047] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159624, 'name': ReconfigVM_Task, 'duration_secs': 0.303569} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.460223] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Reconfigured VM instance instance-0000002d to attach disk [datastore2] c1a6ffac-99bd-492c-99e5-f7c46b352d8e/c1a6ffac-99bd-492c-99e5-f7c46b352d8e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.460837] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ba2934b-9db5-47c6-8c63-3e9476f285bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.464142] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Created folder: Project (a1625ea8d35c4e1ea7291998932e6b17) in parent group-v252271. [ 808.464422] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Creating folder: Instances. Parent ref: group-v252312. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.465072] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df44d7de-2564-4510-b6d3-e0c2154e0ab1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.471667] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 808.471667] env[61986]: value = "task-1159626" [ 808.471667] env[61986]: _type = "Task" [ 808.471667] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.478840] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Created folder: Instances in parent group-v252312. [ 808.479109] env[61986]: DEBUG oslo.service.loopingcall [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.479312] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 808.479525] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48e16623-7ecd-4c23-906c-903660781698 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.498163] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159626, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.504740] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.504740] env[61986]: value = "task-1159628" [ 808.504740] env[61986]: _type = "Task" [ 808.504740] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.513222] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.532288] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159623, 'name': PowerOffVM_Task, 'duration_secs': 0.513511} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.532654] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 808.532907] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 808.533274] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e0c5132-715e-45b7-9e4f-6d8ff4ee554a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.595633] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 808.619079] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.053s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.622018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.126s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.625311] env[61986]: INFO nova.compute.claims [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.631277] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.632060] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.632060] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.632060] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.632378] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.632378] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.632554] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.632765] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.633025] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.633254] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.633525] env[61986]: DEBUG nova.virt.hardware [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.634489] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff34e9ec-bcf4-4a59-9dba-31372157de6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.638808] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 808.639036] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 808.639280] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Deleting the datastore file [datastore2] 4bb588b3-c5e1-4619-b77b-6734d6516170 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.639957] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55dc7ac1-ab33-4550-9a37-59658a7f0587 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.646120] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0ad52a-3528-46f9-927b-3bd8e52ae8e0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.651429] env[61986]: INFO nova.scheduler.client.report [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted allocations for instance 2377b482-52e7-4054-8d5b-8f00587c3371 [ 808.654302] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for the task: (returnval){ [ 808.654302] env[61986]: value = "task-1159630" [ 808.654302] env[61986]: _type = "Task" [ 808.654302] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.681840] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.700867] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e29d25-73f3-a938-f492-146d4b99829b, 'name': SearchDatastore_Task, 'duration_secs': 0.016146} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.702261] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e88474b3-25f5-40ff-8fba-2d9477386a46 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.709064] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 808.709064] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520cce68-d5ea-ae83-2354-6266415dcd5a" [ 808.709064] env[61986]: _type = "Task" [ 808.709064] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.718717] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520cce68-d5ea-ae83-2354-6266415dcd5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.977894] env[61986]: DEBUG nova.network.neutron [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Successfully updated port: 86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.986103] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159626, 'name': Rename_Task, 'duration_secs': 0.214485} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.986560] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.986962] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7d6decc-d433-4e2d-b10c-bf7da06a90fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.995406] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 808.995406] env[61986]: value = "task-1159631" [ 808.995406] env[61986]: _type = "Task" [ 808.995406] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.006416] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.016814] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.166376] env[61986]: DEBUG oslo_concurrency.lockutils [None req-377724e4-b660-4efb-ba15-45db5de792a2 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "2377b482-52e7-4054-8d5b-8f00587c3371" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.991s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.173545] env[61986]: DEBUG oslo_vmware.api [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Task: {'id': task-1159630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294094} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.174490] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.174818] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 809.175219] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 809.175985] env[61986]: INFO nova.compute.manager [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Took 1.68 seconds to destroy the instance on the hypervisor. [ 809.176390] env[61986]: DEBUG oslo.service.loopingcall [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.176729] env[61986]: DEBUG nova.compute.manager [-] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.176930] env[61986]: DEBUG nova.network.neutron [-] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 809.226762] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520cce68-d5ea-ae83-2354-6266415dcd5a, 'name': SearchDatastore_Task, 'duration_secs': 0.0306} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.227530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.228409] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] ce09db77-1eac-4228-a4f8-228a3b7e69d9/ce09db77-1eac-4228-a4f8-228a3b7e69d9.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 809.229048] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-709f9030-cb73-4092-bfed-9b4d07cd8b73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.238634] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 809.238634] env[61986]: value = "task-1159632" [ 809.238634] env[61986]: _type = "Task" [ 809.238634] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.249897] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.482610] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-eaa61f81-2f9b-4d1c-bab8-4363fb71e936" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.482610] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-eaa61f81-2f9b-4d1c-bab8-4363fb71e936" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.482610] env[61986]: DEBUG nova.network.neutron [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.507560] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159631, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.518695] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.526018] env[61986]: DEBUG nova.network.neutron [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Updated VIF entry in instance network info cache for port f1530791-8284-4c8e-a87a-44b3e6145516. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.526388] env[61986]: DEBUG nova.network.neutron [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Updating instance_info_cache with network_info: [{"id": "f1530791-8284-4c8e-a87a-44b3e6145516", "address": "fa:16:3e:e0:69:52", "network": {"id": "ddf922cf-f9d3-449b-b44c-4e89b9f8a306", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1883633649-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a1625ea8d35c4e1ea7291998932e6b17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1530791-82", "ovs_interfaceid": "f1530791-8284-4c8e-a87a-44b3e6145516", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.751687] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.977167] env[61986]: DEBUG nova.network.neutron [-] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.011045] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159631, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.018698] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.028964] env[61986]: DEBUG oslo_concurrency.lockutils [req-61c4b8b9-e452-4d6e-8065-ecd840a61a32 req-09e7432e-a893-4b55-b520-68821ddcda0f service nova] Releasing lock "refresh_cache-8438d2a7-975c-4d1f-8906-f960bfadfe39" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.036095] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf6ea77-8f07-47a3-8990-f150e2fabad6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.045425] env[61986]: DEBUG nova.network.neutron [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.048533] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdfd727-8e13-4bc3-8f8b-b54ddd6a3ac0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.086476] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8be0312-7eb7-4d07-8eaf-aca0edc46d35 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.096506] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1423ba6f-744f-434a-bdc8-3e815a860901 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.113234] env[61986]: DEBUG nova.compute.provider_tree [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.255191] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159632, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.285589] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "98337bb6-9502-4d4c-af00-028659b246bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.285908] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "98337bb6-9502-4d4c-af00-028659b246bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.289585] env[61986]: DEBUG nova.compute.manager [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Received event network-vif-plugged-86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.289585] env[61986]: DEBUG oslo_concurrency.lockutils [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] Acquiring lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.289728] env[61986]: DEBUG oslo_concurrency.lockutils [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.289824] env[61986]: DEBUG oslo_concurrency.lockutils [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.290011] env[61986]: DEBUG nova.compute.manager [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] No waiting events found dispatching network-vif-plugged-86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 810.292537] env[61986]: WARNING nova.compute.manager [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Received unexpected event network-vif-plugged-86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 for instance with vm_state building and task_state spawning. [ 810.292537] env[61986]: DEBUG nova.compute.manager [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Received event network-changed-86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.292537] env[61986]: DEBUG nova.compute.manager [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Refreshing instance network info cache due to event network-changed-86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 810.292537] env[61986]: DEBUG oslo_concurrency.lockutils [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] Acquiring lock "refresh_cache-eaa61f81-2f9b-4d1c-bab8-4363fb71e936" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.386382] env[61986]: DEBUG nova.network.neutron [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Updating instance_info_cache with network_info: [{"id": "86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67", "address": "fa:16:3e:28:5c:5a", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86c32f0f-e3", "ovs_interfaceid": "86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.479841] env[61986]: INFO nova.compute.manager [-] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Took 1.30 seconds to deallocate network for instance. [ 810.512290] env[61986]: DEBUG oslo_vmware.api [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159631, 'name': PowerOnVM_Task, 'duration_secs': 1.365652} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.515351] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.518019] env[61986]: INFO nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Took 9.67 seconds to spawn the instance on the hypervisor. [ 810.518019] env[61986]: DEBUG nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.518019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1083ec-151c-44b6-a999-1edbca885f91 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.524923] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.617018] env[61986]: DEBUG nova.scheduler.client.report [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.751634] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159632, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.122372} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.752016] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] ce09db77-1eac-4228-a4f8-228a3b7e69d9/ce09db77-1eac-4228-a4f8-228a3b7e69d9.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 810.752256] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.752507] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df6c5d13-56e1-417f-80ef-676698704189 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.761023] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 810.761023] env[61986]: value = "task-1159633" [ 810.761023] env[61986]: _type = "Task" [ 810.761023] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.770098] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159633, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.889660] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-eaa61f81-2f9b-4d1c-bab8-4363fb71e936" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.889991] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Instance network_info: |[{"id": "86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67", "address": "fa:16:3e:28:5c:5a", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86c32f0f-e3", "ovs_interfaceid": "86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 810.890332] env[61986]: DEBUG oslo_concurrency.lockutils [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] Acquired lock "refresh_cache-eaa61f81-2f9b-4d1c-bab8-4363fb71e936" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.890527] env[61986]: DEBUG nova.network.neutron [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Refreshing network info cache for port 86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.891781] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:5c:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.900566] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating folder: Project (4e9d0c35a4de4f5f9829a7f3c88fde92). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 810.903824] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c0c2ee0-d054-43c4-8883-a0673bee7ee0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.917985] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Created folder: Project (4e9d0c35a4de4f5f9829a7f3c88fde92) in parent group-v252271. [ 810.918226] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating folder: Instances. Parent ref: group-v252315. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 810.918476] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc44d9c7-8633-492c-9bd6-008f2eb90c37 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.930668] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Created folder: Instances in parent group-v252315. [ 810.930796] env[61986]: DEBUG oslo.service.loopingcall [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.931041] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 810.931131] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b47f841-44e0-4b26-b038-0c3bf46efc6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.955167] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.955167] env[61986]: value = "task-1159636" [ 810.955167] env[61986]: _type = "Task" [ 810.955167] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.963911] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159636, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.988838] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.022067] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.039746] env[61986]: INFO nova.compute.manager [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Took 30.71 seconds to build instance. [ 811.123018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.123573] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.126555] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.448s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.128734] env[61986]: INFO nova.compute.claims [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.275073] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159633, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.200411} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.275487] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.276602] env[61986]: DEBUG nova.network.neutron [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Updated VIF entry in instance network info cache for port 86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.277571] env[61986]: DEBUG nova.network.neutron [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Updating instance_info_cache with network_info: [{"id": "86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67", "address": "fa:16:3e:28:5c:5a", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86c32f0f-e3", "ovs_interfaceid": "86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.283018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc73b2c-16d1-423a-94c3-8c6766e47636 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.316278] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] ce09db77-1eac-4228-a4f8-228a3b7e69d9/ce09db77-1eac-4228-a4f8-228a3b7e69d9.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.317226] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9198752f-2b23-4573-849d-19e69a0f92a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.342355] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 811.342355] env[61986]: value = "task-1159637" [ 811.342355] env[61986]: _type = "Task" [ 811.342355] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.353446] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "9b539a94-6c39-46f8-b194-27047245d1f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.353446] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "9b539a94-6c39-46f8-b194-27047245d1f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.356669] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159637, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.380823] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "d72cd38b-ef14-467b-bf53-97d9e66534c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.380934] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.466802] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159636, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.523816] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.542627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f18c184a-ddf0-4f1d-b958-1a83bbbed300 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.564s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.635381] env[61986]: DEBUG nova.compute.utils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.637850] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 811.640489] env[61986]: DEBUG nova.network.neutron [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 811.733899] env[61986]: DEBUG nova.policy [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ea19b4c355d4b2797683d78e82eadea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0110dfcd30de444badf6fe685042ba3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 811.788327] env[61986]: DEBUG oslo_concurrency.lockutils [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] Releasing lock "refresh_cache-eaa61f81-2f9b-4d1c-bab8-4363fb71e936" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.788598] env[61986]: DEBUG nova.compute.manager [req-066f533a-d979-49c6-9c59-fa8c7d149306 req-d2088278-d71f-488c-995e-e03b81c15615 service nova] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Received event network-vif-deleted-8fa7deb8-c0a2-45a5-98f9-1c07972e34f6 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.854947] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159637, 'name': ReconfigVM_Task, 'duration_secs': 0.298289} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.855356] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Reconfigured VM instance instance-0000002e to attach disk [datastore2] ce09db77-1eac-4228-a4f8-228a3b7e69d9/ce09db77-1eac-4228-a4f8-228a3b7e69d9.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.856190] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b80a5a11-734c-4d30-9d08-d2345d4c50f0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.865632] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 811.865632] env[61986]: value = "task-1159638" [ 811.865632] env[61986]: _type = "Task" [ 811.865632] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.877247] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159638, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.967323] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159636, 'name': CreateVM_Task, 'duration_secs': 0.78597} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.967497] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 811.968191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.968341] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.968668] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.968919] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7600ffe-6ee1-44ff-bf6a-c3d7391cfc13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.974836] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 811.974836] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529ebd2e-ba72-7357-b5ec-4c8dfcfb0701" [ 811.974836] env[61986]: _type = "Task" [ 811.974836] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.983798] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529ebd2e-ba72-7357-b5ec-4c8dfcfb0701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.023087] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.047027] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.138075] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.273976] env[61986]: DEBUG nova.network.neutron [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Successfully created port: d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.379891] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159638, 'name': Rename_Task, 'duration_secs': 0.151311} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.380200] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 812.380421] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a31455e2-7710-4d33-95d9-86a416cdd08f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.387721] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 812.387721] env[61986]: value = "task-1159639" [ 812.387721] env[61986]: _type = "Task" [ 812.387721] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.405518] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159639, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.488369] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529ebd2e-ba72-7357-b5ec-4c8dfcfb0701, 'name': SearchDatastore_Task, 'duration_secs': 0.012872} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.488676] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.488919] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.489141] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.491017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.491017] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.493072] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f95cea2-81e2-49c0-a2b8-c777dc8a82e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.508637] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.508718] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 812.512100] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8fd5666-1adb-4106-96cc-f513dbb6e837 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.523093] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 812.523093] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526cfd42-f11e-9985-74c3-eb3a7ab78664" [ 812.523093] env[61986]: _type = "Task" [ 812.523093] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.526520] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159628, 'name': CreateVM_Task, 'duration_secs': 3.699167} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.532063] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 812.533045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.533138] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.533436] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.533992] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40452fc6-680e-4e49-b3e6-b752c4cc602d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.539792] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526cfd42-f11e-9985-74c3-eb3a7ab78664, 'name': SearchDatastore_Task, 'duration_secs': 0.010509} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.541539] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 812.541539] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5237ef27-04bb-417f-b9a9-c759f2ba1a49" [ 812.541539] env[61986]: _type = "Task" [ 812.541539] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.541731] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a6844a9-4ff0-48de-b0af-733cc20833b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.557676] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 812.557676] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524259e1-ecbd-2117-be84-aab64f3ca417" [ 812.557676] env[61986]: _type = "Task" [ 812.557676] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.568319] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5237ef27-04bb-417f-b9a9-c759f2ba1a49, 'name': SearchDatastore_Task, 'duration_secs': 0.0111} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.571453] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.571776] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.571907] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.576786] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524259e1-ecbd-2117-be84-aab64f3ca417, 'name': SearchDatastore_Task, 'duration_secs': 0.011085} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.577034] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.577300] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] eaa61f81-2f9b-4d1c-bab8-4363fb71e936/eaa61f81-2f9b-4d1c-bab8-4363fb71e936.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.577570] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.577746] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.577947] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87ff1dda-e2ba-459d-af00-422e956f2e67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.580086] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83d2d15d-9391-46b9-a196-b83ae23edf61 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.583308] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.591806] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 812.591806] env[61986]: value = "task-1159640" [ 812.591806] env[61986]: _type = "Task" [ 812.591806] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.593129] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.593309] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 812.596762] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4da7f3c3-2f30-487b-9e58-3390778b47f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.603495] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 812.603495] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520782c3-9474-d584-3ad2-9017aa5521b7" [ 812.603495] env[61986]: _type = "Task" [ 812.603495] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.608186] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.620509] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520782c3-9474-d584-3ad2-9017aa5521b7, 'name': SearchDatastore_Task, 'duration_secs': 0.011432} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.621548] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7202ca67-157e-4644-baa8-8c26812f2563 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.630023] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 812.630023] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aedb38-7768-a50c-ce07-0d64b8164b3d" [ 812.630023] env[61986]: _type = "Task" [ 812.630023] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.641635] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aedb38-7768-a50c-ce07-0d64b8164b3d, 'name': SearchDatastore_Task, 'duration_secs': 0.009473} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.644267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.644559] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 8438d2a7-975c-4d1f-8906-f960bfadfe39/8438d2a7-975c-4d1f-8906-f960bfadfe39.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.649876] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be70c27c-b9e8-4ea7-85a7-54d4edbc0710 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.660931] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 812.660931] env[61986]: value = "task-1159641" [ 812.660931] env[61986]: _type = "Task" [ 812.660931] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.673271] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.712622] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff390ed9-9a78-4289-9c50-61a9f09a7955 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.721424] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81179fb1-029a-4ce1-8353-a4d62096eff0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.753318] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb22289-1326-49c3-86e3-43890adafb31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.764019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275adb6f-87fc-4192-b070-5e853e29e035 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.780109] env[61986]: DEBUG nova.compute.provider_tree [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.900422] env[61986]: DEBUG oslo_vmware.api [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159639, 'name': PowerOnVM_Task, 'duration_secs': 0.504689} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.900698] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 812.900902] env[61986]: INFO nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Took 9.50 seconds to spawn the instance on the hypervisor. [ 812.901206] env[61986]: DEBUG nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.901928] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1e7e18-6def-4877-a722-931cd423cd80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.105036] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159640, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.152454] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.171039] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159641, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.180493] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.180772] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.180930] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.181115] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.181251] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.181393] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.181592] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.181742] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.181899] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.182069] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.182267] env[61986]: DEBUG nova.virt.hardware [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.183183] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cd7741-c6c0-4bc0-a0cb-8f8541a3fb6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.191549] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a22dae-938b-4802-8b6a-273d4233a8df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.283936] env[61986]: DEBUG nova.scheduler.client.report [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.420201] env[61986]: INFO nova.compute.manager [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Took 31.96 seconds to build instance. [ 813.604551] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571228} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.604820] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] eaa61f81-2f9b-4d1c-bab8-4363fb71e936/eaa61f81-2f9b-4d1c-bab8-4363fb71e936.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.605310] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.605394] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f36711e0-2397-4265-983c-62cf3467f19a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.613104] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 813.613104] env[61986]: value = "task-1159642" [ 813.613104] env[61986]: _type = "Task" [ 813.613104] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.623068] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.671895] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.934773} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.672218] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 8438d2a7-975c-4d1f-8906-f960bfadfe39/8438d2a7-975c-4d1f-8906-f960bfadfe39.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.673027] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.673027] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e4222b6-23d4-43af-8bcd-c465de5ed0ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.681188] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 813.681188] env[61986]: value = "task-1159643" [ 813.681188] env[61986]: _type = "Task" [ 813.681188] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.691902] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.789691] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.663s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.790233] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.795013] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.728s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.795261] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.795431] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 813.795718] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.368s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.797334] env[61986]: INFO nova.compute.claims [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.804727] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612b051c-a966-46a8-b25e-286ef0216512 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.816261] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b81289-3be4-4a4c-a7e1-c4a7b61d9f9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.832924] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4c1d04-b649-4da2-aa0f-642561986799 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.841281] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b35d87-d2e9-4c55-bc4c-fac90b11cbb9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.874944] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181362MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 813.875216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.922654] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f056bfb-d657-4701-b123-c5d30ddc11be tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.655s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.126526] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109335} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.126702] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.127551] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c733d16f-2f37-41cf-8365-72e3ae7e9c18 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.151237] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] eaa61f81-2f9b-4d1c-bab8-4363fb71e936/eaa61f81-2f9b-4d1c-bab8-4363fb71e936.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.151679] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6998db88-2b68-4400-bb02-5624db7cd9fd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.174018] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 814.174018] env[61986]: value = "task-1159644" [ 814.174018] env[61986]: _type = "Task" [ 814.174018] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.182967] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159644, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.190971] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.272158] env[61986]: DEBUG nova.compute.manager [req-be316ebf-fa35-4bd0-9a95-2509efeb6270 req-9aba283d-717e-42c6-9108-5900056b213b service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Received event network-vif-plugged-d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.272426] env[61986]: DEBUG oslo_concurrency.lockutils [req-be316ebf-fa35-4bd0-9a95-2509efeb6270 req-9aba283d-717e-42c6-9108-5900056b213b service nova] Acquiring lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.272639] env[61986]: DEBUG oslo_concurrency.lockutils [req-be316ebf-fa35-4bd0-9a95-2509efeb6270 req-9aba283d-717e-42c6-9108-5900056b213b service nova] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.272841] env[61986]: DEBUG oslo_concurrency.lockutils [req-be316ebf-fa35-4bd0-9a95-2509efeb6270 req-9aba283d-717e-42c6-9108-5900056b213b service nova] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.272952] env[61986]: DEBUG nova.compute.manager [req-be316ebf-fa35-4bd0-9a95-2509efeb6270 req-9aba283d-717e-42c6-9108-5900056b213b service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] No waiting events found dispatching network-vif-plugged-d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.273312] env[61986]: WARNING nova.compute.manager [req-be316ebf-fa35-4bd0-9a95-2509efeb6270 req-9aba283d-717e-42c6-9108-5900056b213b service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Received unexpected event network-vif-plugged-d269a844-d4bf-4b1c-bde6-bb3011fa2d06 for instance with vm_state building and task_state spawning. [ 814.309658] env[61986]: DEBUG nova.compute.utils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.315150] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.315310] env[61986]: DEBUG nova.network.neutron [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 814.359682] env[61986]: DEBUG nova.policy [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2377366c4af1411198927bf2ed69f4ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37fa5f7485764468bb537adc78a8177a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 814.362475] env[61986]: DEBUG nova.network.neutron [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Successfully updated port: d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.426091] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.629175] env[61986]: DEBUG nova.network.neutron [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Successfully created port: 5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.684690] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.693537] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.814014] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.822256] env[61986]: DEBUG nova.compute.manager [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.823151] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ea1d18-d3b7-4578-b604-4b02fd0d7826 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.866897] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-6f20dd8e-9289-458d-bed6-f2ef9daaa917" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.866897] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-6f20dd8e-9289-458d-bed6-f2ef9daaa917" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.866897] env[61986]: DEBUG nova.network.neutron [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.945739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.187883] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159644, 'name': ReconfigVM_Task, 'duration_secs': 0.924716} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.191137] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Reconfigured VM instance instance-00000030 to attach disk [datastore1] eaa61f81-2f9b-4d1c-bab8-4363fb71e936/eaa61f81-2f9b-4d1c-bab8-4363fb71e936.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.193953] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ebe7928-f3ef-43d4-bc58-c26437761be7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.201572] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.313894} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.202837] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.203201] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 815.203201] env[61986]: value = "task-1159645" [ 815.203201] env[61986]: _type = "Task" [ 815.203201] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.203889] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e47ff8a-9328-489b-9c1a-1882b3d44a04 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.218204] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159645, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.235737] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 8438d2a7-975c-4d1f-8906-f960bfadfe39/8438d2a7-975c-4d1f-8906-f960bfadfe39.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.237149] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0854bca0-3c66-4dcf-be2a-6ec9427cbf02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.252617] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5d48d1-f933-48ea-b57f-da528cc9cd93 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.262842] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83642bb-a8d6-45a8-be5d-f746355f80ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.265686] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 815.265686] env[61986]: value = "task-1159646" [ 815.265686] env[61986]: _type = "Task" [ 815.265686] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.312942] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b133650-4c98-4f56-bc46-478ee8c89279 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.316743] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159646, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.328931] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb508c2-e8b2-4c61-b0e3-965d612fb26b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.347868] env[61986]: INFO nova.compute.manager [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] instance snapshotting [ 815.350135] env[61986]: DEBUG nova.compute.provider_tree [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.352615] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9278cab3-e36e-4b26-9bde-053a67b29982 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.381760] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563b390b-d943-4534-b54d-b72d10af6ce2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.438488] env[61986]: DEBUG nova.network.neutron [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.716905] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159645, 'name': Rename_Task, 'duration_secs': 0.158442} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.717230] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 815.717457] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bc3f5e3-e467-440f-be6a-01583a700082 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.724820] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 815.724820] env[61986]: value = "task-1159647" [ 815.724820] env[61986]: _type = "Task" [ 815.724820] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.733358] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.776516] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159646, 'name': ReconfigVM_Task, 'duration_secs': 0.331955} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.776824] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 8438d2a7-975c-4d1f-8906-f960bfadfe39/8438d2a7-975c-4d1f-8906-f960bfadfe39.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.777502] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32860afa-ea38-46a2-9685-68189c8aee56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.784833] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 815.784833] env[61986]: value = "task-1159648" [ 815.784833] env[61986]: _type = "Task" [ 815.784833] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.794268] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159648, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.816427] env[61986]: DEBUG nova.network.neutron [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Updating instance_info_cache with network_info: [{"id": "d269a844-d4bf-4b1c-bde6-bb3011fa2d06", "address": "fa:16:3e:b0:5a:35", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd269a844-d4", "ovs_interfaceid": "d269a844-d4bf-4b1c-bde6-bb3011fa2d06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.835498] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.853200] env[61986]: DEBUG nova.scheduler.client.report [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.864961] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.865219] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.865419] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.865608] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.865768] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.865902] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.866125] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.866287] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.866447] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.866663] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.866767] env[61986]: DEBUG nova.virt.hardware [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.867966] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ebe708-afff-4cc0-a8d3-45ef9da1afcb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.877658] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbb8297-bdb1-4300-a798-6e922d3e1bf5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.896157] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 815.896977] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f60ccc65-7a9b-49d5-8fea-dc6110083952 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.905776] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 815.905776] env[61986]: value = "task-1159649" [ 815.905776] env[61986]: _type = "Task" [ 815.905776] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.915943] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159649, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.154447] env[61986]: DEBUG nova.network.neutron [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Successfully updated port: 5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.236456] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159647, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.297245] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159648, 'name': Rename_Task, 'duration_secs': 0.145709} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.297673] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 816.297893] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82045888-503f-4989-a83f-e697a6bad6d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.307452] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 816.307452] env[61986]: value = "task-1159650" [ 816.307452] env[61986]: _type = "Task" [ 816.307452] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.316893] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.323598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-6f20dd8e-9289-458d-bed6-f2ef9daaa917" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.323927] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance network_info: |[{"id": "d269a844-d4bf-4b1c-bde6-bb3011fa2d06", "address": "fa:16:3e:b0:5a:35", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd269a844-d4", "ovs_interfaceid": "d269a844-d4bf-4b1c-bde6-bb3011fa2d06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.324364] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:5a:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd269a844-d4bf-4b1c-bde6-bb3011fa2d06', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.333335] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating folder: Project (0110dfcd30de444badf6fe685042ba3b). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.333554] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bda39f38-2d0c-46be-8fd5-31a26fecb5b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.347390] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created folder: Project (0110dfcd30de444badf6fe685042ba3b) in parent group-v252271. [ 816.347637] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating folder: Instances. Parent ref: group-v252318. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.347910] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e96f657-821c-4386-83f0-6d460e775947 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.360262] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.360937] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.366489] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 14.290s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.368752] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created folder: Instances in parent group-v252318. [ 816.369800] env[61986]: DEBUG oslo.service.loopingcall [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.369800] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.369908] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ea89a8f-6254-45fd-8092-da4050b1959a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.397922] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.397922] env[61986]: value = "task-1159653" [ 816.397922] env[61986]: _type = "Task" [ 816.397922] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.408249] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159653, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.417743] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159649, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.657045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "refresh_cache-ebf5fee5-7f1b-4537-aec3-77a8a963670b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.657493] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquired lock "refresh_cache-ebf5fee5-7f1b-4537-aec3-77a8a963670b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.657824] env[61986]: DEBUG nova.network.neutron [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.685522] env[61986]: DEBUG nova.compute.manager [req-e0361c2a-52f5-48fc-a885-65f08f5c957f req-3e25b666-d5b3-4d16-bea6-9efdaa099394 service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Received event network-vif-plugged-5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.685522] env[61986]: DEBUG oslo_concurrency.lockutils [req-e0361c2a-52f5-48fc-a885-65f08f5c957f req-3e25b666-d5b3-4d16-bea6-9efdaa099394 service nova] Acquiring lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.687141] env[61986]: DEBUG oslo_concurrency.lockutils [req-e0361c2a-52f5-48fc-a885-65f08f5c957f req-3e25b666-d5b3-4d16-bea6-9efdaa099394 service nova] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.687141] env[61986]: DEBUG oslo_concurrency.lockutils [req-e0361c2a-52f5-48fc-a885-65f08f5c957f req-3e25b666-d5b3-4d16-bea6-9efdaa099394 service nova] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.687141] env[61986]: DEBUG nova.compute.manager [req-e0361c2a-52f5-48fc-a885-65f08f5c957f req-3e25b666-d5b3-4d16-bea6-9efdaa099394 service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] No waiting events found dispatching network-vif-plugged-5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.687141] env[61986]: WARNING nova.compute.manager [req-e0361c2a-52f5-48fc-a885-65f08f5c957f req-3e25b666-d5b3-4d16-bea6-9efdaa099394 service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Received unexpected event network-vif-plugged-5ee8af5d-5341-4ee4-9245-cfa7dc703dca for instance with vm_state building and task_state spawning. [ 816.737173] env[61986]: DEBUG oslo_vmware.api [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159647, 'name': PowerOnVM_Task, 'duration_secs': 0.688654} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.738049] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 816.738353] env[61986]: INFO nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Took 8.14 seconds to spawn the instance on the hypervisor. [ 816.738604] env[61986]: DEBUG nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.742956] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab4d4a5-86fe-48b5-a666-64eaa5bbd514 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.824769] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159650, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.871499] env[61986]: DEBUG nova.compute.utils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.876377] env[61986]: INFO nova.compute.claims [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.880226] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.880289] env[61986]: DEBUG nova.network.neutron [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.910412] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159653, 'name': CreateVM_Task, 'duration_secs': 0.445575} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.913394] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 816.914749] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.914909] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.915310] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.916165] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e9d8f97-1423-4757-af61-ed3aa0dc54ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.921862] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159649, 'name': CreateSnapshot_Task, 'duration_secs': 0.750008} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.923401] env[61986]: DEBUG nova.policy [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c93b2022a554633b279dba6354f23f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50561f3dfae840348c86bcce568d354f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 816.926053] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 816.926689] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 816.926689] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5280fd43-d3e7-8ebe-ec27-8bf3fed9ec38" [ 816.926689] env[61986]: _type = "Task" [ 816.926689] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.927458] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f50a594-ff64-45a0-b2f6-031a7906f403 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.935730] env[61986]: DEBUG nova.compute.manager [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Received event network-changed-d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.935916] env[61986]: DEBUG nova.compute.manager [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Refreshing instance network info cache due to event network-changed-d269a844-d4bf-4b1c-bde6-bb3011fa2d06. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 816.936139] env[61986]: DEBUG oslo_concurrency.lockutils [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] Acquiring lock "refresh_cache-6f20dd8e-9289-458d-bed6-f2ef9daaa917" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.936277] env[61986]: DEBUG oslo_concurrency.lockutils [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] Acquired lock "refresh_cache-6f20dd8e-9289-458d-bed6-f2ef9daaa917" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.936458] env[61986]: DEBUG nova.network.neutron [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Refreshing network info cache for port d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.947473] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5280fd43-d3e7-8ebe-ec27-8bf3fed9ec38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.199125] env[61986]: DEBUG nova.network.neutron [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Successfully created port: 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.217046] env[61986]: DEBUG nova.network.neutron [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.262826] env[61986]: INFO nova.compute.manager [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Took 27.17 seconds to build instance. [ 817.322881] env[61986]: DEBUG oslo_vmware.api [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159650, 'name': PowerOnVM_Task, 'duration_secs': 0.552296} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.323571] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 817.323971] env[61986]: INFO nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Took 11.32 seconds to spawn the instance on the hypervisor. [ 817.323971] env[61986]: DEBUG nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.326183] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a330fa8d-a2e1-4b39-ac39-e712554d960f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.380927] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.392178] env[61986]: INFO nova.compute.resource_tracker [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating resource usage from migration 84ab935e-920f-44a1-8d8e-f5b0438906b8 [ 817.450030] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5280fd43-d3e7-8ebe-ec27-8bf3fed9ec38, 'name': SearchDatastore_Task, 'duration_secs': 0.028394} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.452145] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.452517] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.452872] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.453103] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.453344] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.462164] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 817.465259] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfe64d17-ecd6-4a3a-a884-1b51450cde24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.471044] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-58ef03cc-d60c-4eac-baa9-6c0fc385d54b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.483455] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 817.483455] env[61986]: value = "task-1159654" [ 817.483455] env[61986]: _type = "Task" [ 817.483455] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.484717] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.484895] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.488757] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e04442a7-0038-40ca-a2df-cc1e7ce93ae9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.498861] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 817.498861] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520efb00-af43-e388-57ab-b5322b402df2" [ 817.498861] env[61986]: _type = "Task" [ 817.498861] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.502724] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159654, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.517140] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520efb00-af43-e388-57ab-b5322b402df2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.712773] env[61986]: DEBUG nova.network.neutron [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Updated VIF entry in instance network info cache for port d269a844-d4bf-4b1c-bde6-bb3011fa2d06. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.713220] env[61986]: DEBUG nova.network.neutron [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Updating instance_info_cache with network_info: [{"id": "d269a844-d4bf-4b1c-bde6-bb3011fa2d06", "address": "fa:16:3e:b0:5a:35", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd269a844-d4", "ovs_interfaceid": "d269a844-d4bf-4b1c-bde6-bb3011fa2d06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.753063] env[61986]: DEBUG nova.network.neutron [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Updating instance_info_cache with network_info: [{"id": "5ee8af5d-5341-4ee4-9245-cfa7dc703dca", "address": "fa:16:3e:48:3e:be", "network": {"id": "e17d78b2-6057-4cff-987b-aca7b7e2479e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-554396804-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fa5f7485764468bb537adc78a8177a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ee8af5d-53", "ovs_interfaceid": "5ee8af5d-5341-4ee4-9245-cfa7dc703dca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.764175] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e651d4-0f9f-4036-a124-e9b7b7021362 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.336s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.857000] env[61986]: INFO nova.compute.manager [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Took 29.50 seconds to build instance. [ 817.920342] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d699c7-e3f5-4725-80b1-7193df103b6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.929860] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe434fd-f177-4246-a4ef-ae4cdf977b77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.965900] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedafa6f-09dd-4739-89bc-8f947e66ff44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.975432] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6682d01-4355-4f4e-a7ec-dccc707dde8e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.991209] env[61986]: DEBUG nova.compute.provider_tree [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.004365] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159654, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.014985] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520efb00-af43-e388-57ab-b5322b402df2, 'name': SearchDatastore_Task, 'duration_secs': 0.034456} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.016128] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5222664-7859-4cdf-b581-6d6b40c3aade {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.022187] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 818.022187] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ddb49d-43f5-e9f3-a98e-7afc9cb24a46" [ 818.022187] env[61986]: _type = "Task" [ 818.022187] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.031978] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ddb49d-43f5-e9f3-a98e-7afc9cb24a46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.218679] env[61986]: DEBUG oslo_concurrency.lockutils [req-6a6e1544-ddb4-4002-bdea-d98613cdc7a9 req-43fd76df-bfe6-4b0c-9326-dfa201590518 service nova] Releasing lock "refresh_cache-6f20dd8e-9289-458d-bed6-f2ef9daaa917" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.256693] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Releasing lock "refresh_cache-ebf5fee5-7f1b-4537-aec3-77a8a963670b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.257072] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Instance network_info: |[{"id": "5ee8af5d-5341-4ee4-9245-cfa7dc703dca", "address": "fa:16:3e:48:3e:be", "network": {"id": "e17d78b2-6057-4cff-987b-aca7b7e2479e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-554396804-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fa5f7485764468bb537adc78a8177a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ee8af5d-53", "ovs_interfaceid": "5ee8af5d-5341-4ee4-9245-cfa7dc703dca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.257679] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:3e:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd47d5e1d-e66d-4f2c-83e6-d5e78c2b767d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ee8af5d-5341-4ee4-9245-cfa7dc703dca', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.266507] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Creating folder: Project (37fa5f7485764468bb537adc78a8177a). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.266966] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.269908] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9774425-631f-4ab8-afd7-007aa4a34033 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.286409] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Created folder: Project (37fa5f7485764468bb537adc78a8177a) in parent group-v252271. [ 818.286780] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Creating folder: Instances. Parent ref: group-v252323. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.287182] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c2ab156-e285-42e5-b306-e860ec1bdf96 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.302226] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Created folder: Instances in parent group-v252323. [ 818.302741] env[61986]: DEBUG oslo.service.loopingcall [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.303013] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 818.303362] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9518680-0635-4ba7-8e5b-5b201fa6129b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.327416] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.327416] env[61986]: value = "task-1159657" [ 818.327416] env[61986]: _type = "Task" [ 818.327416] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.337055] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159657, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.359742] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24719d8e-ff3a-470b-b021-edbcc1222828 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.218s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.401971] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.437210] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.437548] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.437706] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.437883] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.439637] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.439937] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.440262] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.440447] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.440638] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.440864] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.441091] env[61986]: DEBUG nova.virt.hardware [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.442435] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4406128-e1d3-44c4-aa1b-8d4768a827fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.451980] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f485487a-b6b5-47ca-8bb7-ee69b3afa1ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.497543] env[61986]: DEBUG nova.scheduler.client.report [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.506663] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159654, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.536193] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ddb49d-43f5-e9f3-a98e-7afc9cb24a46, 'name': SearchDatastore_Task, 'duration_secs': 0.011234} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.536497] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.536741] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 818.537051] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-66a2f9f9-65ba-4391-9089-80e6c4ffcf7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.570169] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 818.570169] env[61986]: value = "task-1159658" [ 818.570169] env[61986]: _type = "Task" [ 818.570169] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.581786] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.756416] env[61986]: DEBUG nova.network.neutron [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Successfully updated port: 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.800640] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.847136] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159657, 'name': CreateVM_Task, 'duration_secs': 0.368036} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.847387] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 818.848717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.848717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.848717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 818.848962] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3e6995c-ac03-4493-93f7-5cf85b1edde6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.857787] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 818.857787] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5205ad7b-aeb2-fa91-6d12-f70800767a29" [ 818.857787] env[61986]: _type = "Task" [ 818.857787] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.865141] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.876233] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5205ad7b-aeb2-fa91-6d12-f70800767a29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.007482] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159654, 'name': CloneVM_Task, 'duration_secs': 1.224431} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.007482] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Created linked-clone VM from snapshot [ 819.007482] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136326eb-29ca-48b4-a896-bc106f94c49e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.010507] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.643s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.010878] env[61986]: INFO nova.compute.manager [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Migrating [ 819.011275] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.011565] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.013211] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.972s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.016530] env[61986]: INFO nova.compute.claims [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.028977] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Uploading image 297420cf-bb6c-4065-9250-42a9ab8f58f1 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 819.043299] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 819.043299] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6d834af5-964a-4d68-b7ed-d1db7ea6d8fc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.053054] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 819.053054] env[61986]: value = "task-1159659" [ 819.053054] env[61986]: _type = "Task" [ 819.053054] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.064436] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159659, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.083883] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159658, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.258809] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.258969] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.259138] env[61986]: DEBUG nova.network.neutron [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 819.291406] env[61986]: DEBUG nova.compute.manager [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Received event network-changed-5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.292203] env[61986]: DEBUG nova.compute.manager [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Refreshing instance network info cache due to event network-changed-5ee8af5d-5341-4ee4-9245-cfa7dc703dca. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 819.292203] env[61986]: DEBUG oslo_concurrency.lockutils [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] Acquiring lock "refresh_cache-ebf5fee5-7f1b-4537-aec3-77a8a963670b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.292313] env[61986]: DEBUG oslo_concurrency.lockutils [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] Acquired lock "refresh_cache-ebf5fee5-7f1b-4537-aec3-77a8a963670b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.292520] env[61986]: DEBUG nova.network.neutron [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Refreshing network info cache for port 5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.319054] env[61986]: DEBUG nova.compute.manager [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Received event network-changed-9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.319331] env[61986]: DEBUG nova.compute.manager [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Refreshing instance network info cache due to event network-changed-9ffa532a-af81-442f-9253-1c78d86bbd6f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 819.319607] env[61986]: DEBUG oslo_concurrency.lockutils [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] Acquiring lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.319800] env[61986]: DEBUG oslo_concurrency.lockutils [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] Acquired lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.320030] env[61986]: DEBUG nova.network.neutron [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Refreshing network info cache for port 9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.373586] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5205ad7b-aeb2-fa91-6d12-f70800767a29, 'name': SearchDatastore_Task, 'duration_secs': 0.060322} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.373890] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.374127] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.374361] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.374515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.374758] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.374961] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfb10e93-af10-4675-92e3-df43b6feb438 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.388676] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.388850] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 819.390996] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b848bff3-5bcc-4c59-8f2e-25d316b46564 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.392825] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.396670] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 819.396670] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a1e137-32b2-8d61-58d1-527e6ab752ce" [ 819.396670] env[61986]: _type = "Task" [ 819.396670] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.405644] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a1e137-32b2-8d61-58d1-527e6ab752ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.521164] env[61986]: INFO nova.compute.rpcapi [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 819.521705] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.566240] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159659, 'name': Destroy_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.579649] env[61986]: DEBUG nova.compute.manager [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.584177] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2f7b9e-a41a-4daf-9e05-d621f8f02bc7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.587076] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159658, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.793140] env[61986]: DEBUG nova.network.neutron [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.909325] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a1e137-32b2-8d61-58d1-527e6ab752ce, 'name': SearchDatastore_Task, 'duration_secs': 0.014534} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.912417] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02f100c6-75d8-48a7-a267-74c2d958eab2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.919053] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 819.919053] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5209aca6-e3dc-8800-d1b6-c44ee3fb743d" [ 819.919053] env[61986]: _type = "Task" [ 819.919053] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.934144] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5209aca6-e3dc-8800-d1b6-c44ee3fb743d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.970705] env[61986]: DEBUG nova.network.neutron [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.022991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "8438d2a7-975c-4d1f-8906-f960bfadfe39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.022991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.022991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "8438d2a7-975c-4d1f-8906-f960bfadfe39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.022991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.023302] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.023879] env[61986]: INFO nova.compute.manager [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Terminating instance [ 820.025919] env[61986]: DEBUG nova.compute.manager [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.026259] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 820.027209] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be28038b-35fa-41de-a8a0-ac02fd6f2f33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.042029] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 820.042943] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.043518] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.043818] env[61986]: DEBUG nova.network.neutron [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.045310] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fc054ee-9f14-4a67-9966-1931a6b0d561 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.056449] env[61986]: DEBUG oslo_vmware.api [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 820.056449] env[61986]: value = "task-1159660" [ 820.056449] env[61986]: _type = "Task" [ 820.056449] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.073488] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159659, 'name': Destroy_Task, 'duration_secs': 0.552707} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.080042] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Destroyed the VM [ 820.080445] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 820.080758] env[61986]: DEBUG oslo_vmware.api [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.084120] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ad049807-3451-4491-84f3-252c710280f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.096038] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 820.096038] env[61986]: value = "task-1159661" [ 820.096038] env[61986]: _type = "Task" [ 820.096038] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.097066] env[61986]: INFO nova.compute.manager [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] instance snapshotting [ 820.098936] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159658, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.508788} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.101880] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.102095] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.102905] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666a7ff9-0994-4196-ae6f-279f5f01bd1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.111257] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1342fe90-c098-4913-95b7-eeda78075008 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.121788] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159661, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.138892] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 820.138892] env[61986]: value = "task-1159662" [ 820.138892] env[61986]: _type = "Task" [ 820.138892] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.144888] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a68210-3f29-4430-ac5a-742eff945478 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.166102] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159662, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.193144] env[61986]: DEBUG nova.network.neutron [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Updated VIF entry in instance network info cache for port 5ee8af5d-5341-4ee4-9245-cfa7dc703dca. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.197031] env[61986]: DEBUG nova.network.neutron [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Updating instance_info_cache with network_info: [{"id": "5ee8af5d-5341-4ee4-9245-cfa7dc703dca", "address": "fa:16:3e:48:3e:be", "network": {"id": "e17d78b2-6057-4cff-987b-aca7b7e2479e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-554396804-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fa5f7485764468bb537adc78a8177a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ee8af5d-53", "ovs_interfaceid": "5ee8af5d-5341-4ee4-9245-cfa7dc703dca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.197031] env[61986]: DEBUG nova.network.neutron [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updated VIF entry in instance network info cache for port 9ffa532a-af81-442f-9253-1c78d86bbd6f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.197518] env[61986]: DEBUG nova.network.neutron [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updating instance_info_cache with network_info: [{"id": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "address": "fa:16:3e:32:23:4b", "network": {"id": "c15ae3c7-4631-4c01-9027-64898fd0163c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1866800563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b5abb4c676745be8d77ad83e7e0a9ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ffa532a-af", "ovs_interfaceid": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.433236] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5209aca6-e3dc-8800-d1b6-c44ee3fb743d, 'name': SearchDatastore_Task, 'duration_secs': 0.053041} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.433502] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.433752] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] ebf5fee5-7f1b-4537-aec3-77a8a963670b/ebf5fee5-7f1b-4537-aec3-77a8a963670b.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 820.434017] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a50ea01a-130b-4f04-8adc-c197b605cbb9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.444316] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 820.444316] env[61986]: value = "task-1159663" [ 820.444316] env[61986]: _type = "Task" [ 820.444316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.455658] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.473705] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.474053] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Instance network_info: |[{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.474540] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:c4:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496ac502-bfc4-4324-8332-cac473eb7cc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '812b5c34-903f-47de-b3f8-649b3a84eb27', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.482808] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Creating folder: Project (50561f3dfae840348c86bcce568d354f). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.483190] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0b4efcf-8730-40a8-9c67-3c2519a1db63 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.501311] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Created folder: Project (50561f3dfae840348c86bcce568d354f) in parent group-v252271. [ 820.501529] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Creating folder: Instances. Parent ref: group-v252326. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.501793] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ff6914c-3590-4dd5-8c2c-f8efd14c4afe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.513346] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Created folder: Instances in parent group-v252326. [ 820.513591] env[61986]: DEBUG oslo.service.loopingcall [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.513780] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 820.514248] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8daf0653-6273-493d-9bad-366142b36d61 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.538974] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.538974] env[61986]: value = "task-1159666" [ 820.538974] env[61986]: _type = "Task" [ 820.538974] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.550569] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159666, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.574109] env[61986]: DEBUG oslo_vmware.api [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159660, 'name': PowerOffVM_Task, 'duration_secs': 0.216136} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.574416] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 820.574631] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 820.575264] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66c328f0-a246-4c00-9ab3-accb7b5b6886 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.579250] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdad2dec-ad25-42cd-b011-32dcf17c3ac6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.587820] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea9001d-6214-4061-9892-1d0cd011e839 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.630638] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04aeb8b3-21d0-4758-b2a2-2885c06bd064 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.643407] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e59315-8708-4c69-8b44-351ad2122fed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.648018] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159661, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.668288] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 820.669056] env[61986]: DEBUG nova.compute.provider_tree [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.673444] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2ddfb2dd-96c8-4852-ad53-6e9d8e712c0d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.675810] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159662, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.18359} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.679378] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.681824] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac228732-11f3-45b8-8d4d-de9079cabb70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.690650] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 820.690650] env[61986]: value = "task-1159668" [ 820.690650] env[61986]: _type = "Task" [ 820.690650] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.702109] env[61986]: DEBUG oslo_concurrency.lockutils [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] Releasing lock "refresh_cache-ebf5fee5-7f1b-4537-aec3-77a8a963670b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.702379] env[61986]: DEBUG nova.compute.manager [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-vif-plugged-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.702578] env[61986]: DEBUG oslo_concurrency.lockutils [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] Acquiring lock "86c9e45c-b3f1-4004-9acc-190b11a4a926-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.702783] env[61986]: DEBUG oslo_concurrency.lockutils [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.702950] env[61986]: DEBUG oslo_concurrency.lockutils [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.703139] env[61986]: DEBUG nova.compute.manager [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] No waiting events found dispatching network-vif-plugged-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.703311] env[61986]: WARNING nova.compute.manager [req-89bc9923-c288-4866-9588-0c6bb39a6987 req-92ce2c82-a810-4e28-a686-95b522468f4c service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received unexpected event network-vif-plugged-812b5c34-903f-47de-b3f8-649b3a84eb27 for instance with vm_state building and task_state spawning. [ 820.704010] env[61986]: DEBUG oslo_concurrency.lockutils [req-3d04b6bd-8671-4c6a-901f-fe77174b548e req-ff7b47fc-ad68-4a2f-9821-3c88d9eb49d1 service nova] Releasing lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.713727] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.718893] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e5824e4-e7fa-4e78-842d-17f5793f4157 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.736163] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 820.736471] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 820.736717] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Deleting the datastore file [datastore1] 8438d2a7-975c-4d1f-8906-f960bfadfe39 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.741093] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c62a83d2-0bf5-45dc-a99c-a624407068ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.753725] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159668, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.756793] env[61986]: DEBUG oslo_vmware.api [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for the task: (returnval){ [ 820.756793] env[61986]: value = "task-1159670" [ 820.756793] env[61986]: _type = "Task" [ 820.756793] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.757168] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 820.757168] env[61986]: value = "task-1159669" [ 820.757168] env[61986]: _type = "Task" [ 820.757168] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.774644] env[61986]: DEBUG oslo_vmware.api [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.779103] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159669, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.894527] env[61986]: DEBUG nova.network.neutron [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.956639] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159663, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.052353] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159666, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.136870] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159661, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.181271] env[61986]: DEBUG nova.scheduler.client.report [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.211945] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159668, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.271834] env[61986]: DEBUG oslo_vmware.api [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Task: {'id': task-1159670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.347803} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.278490] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.278490] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 821.278490] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 821.278490] env[61986]: INFO nova.compute.manager [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Took 1.25 seconds to destroy the instance on the hypervisor. [ 821.278490] env[61986]: DEBUG oslo.service.loopingcall [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.278766] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159669, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.278766] env[61986]: DEBUG nova.compute.manager [-] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.278766] env[61986]: DEBUG nova.network.neutron [-] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 821.397361] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.457242] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515337} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.457541] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] ebf5fee5-7f1b-4537-aec3-77a8a963670b/ebf5fee5-7f1b-4537-aec3-77a8a963670b.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 821.457826] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.458209] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2efa40e8-bf2c-4d85-936c-8dad8a3e8e5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.468027] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 821.468027] env[61986]: value = "task-1159671" [ 821.468027] env[61986]: _type = "Task" [ 821.468027] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.479036] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159671, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.569300] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159666, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.637647] env[61986]: DEBUG oslo_vmware.api [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159661, 'name': RemoveSnapshot_Task, 'duration_secs': 1.452139} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.637923] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 821.651561] env[61986]: DEBUG nova.compute.manager [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.651804] env[61986]: DEBUG nova.compute.manager [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing instance network info cache due to event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.652162] env[61986]: DEBUG oslo_concurrency.lockutils [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] Acquiring lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.652162] env[61986]: DEBUG oslo_concurrency.lockutils [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] Acquired lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.652275] env[61986]: DEBUG nova.network.neutron [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 821.689487] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.690169] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.694189] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.343s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.694189] env[61986]: DEBUG nova.objects.instance [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lazy-loading 'resources' on Instance uuid 6c62a1e9-d60c-4720-a554-951bea4b2e18 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.716185] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159668, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.743942] env[61986]: DEBUG nova.compute.manager [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Received event network-changed-9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.744154] env[61986]: DEBUG nova.compute.manager [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Refreshing instance network info cache due to event network-changed-9ffa532a-af81-442f-9253-1c78d86bbd6f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.744362] env[61986]: DEBUG oslo_concurrency.lockutils [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] Acquiring lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.744576] env[61986]: DEBUG oslo_concurrency.lockutils [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] Acquired lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.744648] env[61986]: DEBUG nova.network.neutron [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Refreshing network info cache for port 9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 821.770829] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159669, 'name': ReconfigVM_Task, 'duration_secs': 0.770248} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.771468] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.772177] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69c1d94e-4e4c-4044-8b45-e11b41f2048b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.782833] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 821.782833] env[61986]: value = "task-1159672" [ 821.782833] env[61986]: _type = "Task" [ 821.782833] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.795355] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159672, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.979161] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159671, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.250249} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.979541] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.980346] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce05e3c-7426-47be-b63d-9528c8170713 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.003663] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] ebf5fee5-7f1b-4537-aec3-77a8a963670b/ebf5fee5-7f1b-4537-aec3-77a8a963670b.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.004744] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5f2f18d-dff5-4109-a464-03eacdc624f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.025931] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 822.025931] env[61986]: value = "task-1159673" [ 822.025931] env[61986]: _type = "Task" [ 822.025931] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.034745] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.050947] env[61986]: DEBUG nova.network.neutron [-] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.052174] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159666, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.145753] env[61986]: WARNING nova.compute.manager [None req-2e1a6020-6847-45bb-b18e-b27596a24640 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Image not found during snapshot: nova.exception.ImageNotFound: Image 297420cf-bb6c-4065-9250-42a9ab8f58f1 could not be found. [ 822.199667] env[61986]: DEBUG nova.compute.utils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.201244] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.201600] env[61986]: DEBUG nova.network.neutron [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 822.213306] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159668, 'name': CreateSnapshot_Task, 'duration_secs': 1.453893} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.214026] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 822.214899] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bd9138-dd41-42ad-9fbb-90cabe89bd6e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.296797] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159672, 'name': Rename_Task, 'duration_secs': 0.338059} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.296797] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.296797] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f2b1256-3849-453d-844b-9ac0d6cd8795 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.303595] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 822.303595] env[61986]: value = "task-1159674" [ 822.303595] env[61986]: _type = "Task" [ 822.303595] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.310336] env[61986]: DEBUG nova.policy [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8562656b1e46628059ea24f9e16b26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0db18c866dc4ffb8bcb050f8ec6021d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 822.326813] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159674, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.535198] env[61986]: DEBUG nova.network.neutron [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updated VIF entry in instance network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.535198] env[61986]: DEBUG nova.network.neutron [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.539544] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159673, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.553413] env[61986]: INFO nova.compute.manager [-] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Took 1.28 seconds to deallocate network for instance. [ 822.576307] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159666, 'name': CreateVM_Task, 'duration_secs': 1.587899} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.576307] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 822.579362] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.579729] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.580393] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.580978] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a1428aa-1fba-4bc5-bca8-2950da2a1159 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.589018] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 822.589018] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5268f21a-842d-4bbf-0a2c-b44dc21c6032" [ 822.589018] env[61986]: _type = "Task" [ 822.589018] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.596438] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5268f21a-842d-4bbf-0a2c-b44dc21c6032, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.620847] env[61986]: DEBUG nova.network.neutron [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updated VIF entry in instance network info cache for port 9ffa532a-af81-442f-9253-1c78d86bbd6f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.621287] env[61986]: DEBUG nova.network.neutron [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updating instance_info_cache with network_info: [{"id": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "address": "fa:16:3e:32:23:4b", "network": {"id": "c15ae3c7-4631-4c01-9027-64898fd0163c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1866800563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5b5abb4c676745be8d77ad83e7e0a9ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ffa532a-af", "ovs_interfaceid": "9ffa532a-af81-442f-9253-1c78d86bbd6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.695153] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b196f7-b9f8-487c-919c-b4f827e3820e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.704814] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4f7cef-dcfb-4d3d-ad2f-db7fc93b2a56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.708545] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.756907] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 822.758734] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9b895e68-813c-45a6-aee9-e65ca12be3bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.763732] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24ca2c7-0da2-4fe2-9190-ae2a3960abed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.777889] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8249cf67-b9ca-4d0b-a83f-6bb264b73bba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.783873] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 822.783873] env[61986]: value = "task-1159675" [ 822.783873] env[61986]: _type = "Task" [ 822.783873] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.804027] env[61986]: DEBUG nova.compute.provider_tree [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.810319] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159675, 'name': CloneVM_Task} progress is 11%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.824193] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159674, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.862817] env[61986]: DEBUG nova.network.neutron [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Successfully created port: e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.917680] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c89b24b-1c2b-4072-80ad-9afb5a865bd4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.938557] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 823.036893] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159673, 'name': ReconfigVM_Task, 'duration_secs': 0.653944} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.037188] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Reconfigured VM instance instance-00000032 to attach disk [datastore1] ebf5fee5-7f1b-4537-aec3-77a8a963670b/ebf5fee5-7f1b-4537-aec3-77a8a963670b.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.037884] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d0dfc7b-8804-4912-85a5-a4ca9a956ec5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.041202] env[61986]: DEBUG oslo_concurrency.lockutils [req-cb45dfca-eb09-440e-8503-5dd67f6fb493 req-95a2fcc3-b13d-456c-a4c1-eb1e81621557 service nova] Releasing lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.045601] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 823.045601] env[61986]: value = "task-1159676" [ 823.045601] env[61986]: _type = "Task" [ 823.045601] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.054189] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159676, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.077013] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.099877] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5268f21a-842d-4bbf-0a2c-b44dc21c6032, 'name': SearchDatastore_Task, 'duration_secs': 0.012961} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.100331] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.100619] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.100949] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.103884] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.104232] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.104638] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8452900-9f82-4881-9e93-705e746c048d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.115209] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.115558] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 823.116379] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ffed48a-cbdb-497c-a2e0-1cd3e3090642 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.122534] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 823.122534] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b9eb85-ae26-8a05-de1d-894a29bbacb4" [ 823.122534] env[61986]: _type = "Task" [ 823.122534] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.127132] env[61986]: DEBUG oslo_concurrency.lockutils [req-cd8ffeb9-84c9-4508-b88e-f63bf5f960c1 req-f7ab9b58-bc2b-4a38-953c-8588d2e3056b service nova] Releasing lock "refresh_cache-ce09db77-1eac-4228-a4f8-228a3b7e69d9" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.134315] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b9eb85-ae26-8a05-de1d-894a29bbacb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.177541] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.177793] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.178142] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.178377] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.178572] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.181304] env[61986]: INFO nova.compute.manager [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Terminating instance [ 823.183479] env[61986]: DEBUG nova.compute.manager [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 823.183710] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 823.184569] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad66e44-c629-4778-8596-977c1093933c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.193073] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 823.193403] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f00e36e-4a63-4a9e-b74b-695a8bb481f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.200923] env[61986]: DEBUG oslo_vmware.api [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 823.200923] env[61986]: value = "task-1159677" [ 823.200923] env[61986]: _type = "Task" [ 823.200923] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.211550] env[61986]: DEBUG oslo_vmware.api [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.297085] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159675, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.305223] env[61986]: DEBUG nova.scheduler.client.report [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.321568] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159674, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.445547] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 823.446087] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b0ca0fc-8436-47e0-a1a4-02a0dbac4d46 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.454770] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 823.454770] env[61986]: value = "task-1159678" [ 823.454770] env[61986]: _type = "Task" [ 823.454770] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.468629] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.557390] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159676, 'name': Rename_Task, 'duration_secs': 0.171563} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.557834] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 823.558203] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78e48ba9-103d-4dab-9317-ad7c9b91d17c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.566452] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 823.566452] env[61986]: value = "task-1159679" [ 823.566452] env[61986]: _type = "Task" [ 823.566452] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.585690] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.635985] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b9eb85-ae26-8a05-de1d-894a29bbacb4, 'name': SearchDatastore_Task, 'duration_secs': 0.01989} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.637813] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836323f6-00b6-4c12-8c06-154c54ec505c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.646022] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 823.646022] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ab2568-2fa9-5e6b-9152-5ca174dfb26b" [ 823.646022] env[61986]: _type = "Task" [ 823.646022] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.653943] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ab2568-2fa9-5e6b-9152-5ca174dfb26b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.712573] env[61986]: DEBUG oslo_vmware.api [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159677, 'name': PowerOffVM_Task, 'duration_secs': 0.347585} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.712573] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 823.712573] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 823.712573] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4983d79-39af-4a5c-a4ed-22b7b48f96c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.720027] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.751304] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.751545] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.751736] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.751940] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.752095] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.752244] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.752449] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.752603] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.752765] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.752921] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.753105] env[61986]: DEBUG nova.virt.hardware [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.753981] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8d1749-4e15-4427-abd4-3794480c2490 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.763415] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596146fe-cb1c-4569-a342-63d9f47dab53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.782219] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 823.782439] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 823.782616] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleting the datastore file [datastore2] c1a6ffac-99bd-492c-99e5-f7c46b352d8e {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.782869] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee8f593a-e82d-4b46-992e-5ce9dba33ef9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.790490] env[61986]: DEBUG oslo_vmware.api [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 823.790490] env[61986]: value = "task-1159681" [ 823.790490] env[61986]: _type = "Task" [ 823.790490] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.794179] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159675, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.802300] env[61986]: DEBUG oslo_vmware.api [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.812260] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.118s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.815021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.413s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.816903] env[61986]: INFO nova.compute.claims [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.830768] env[61986]: DEBUG oslo_vmware.api [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159674, 'name': PowerOnVM_Task, 'duration_secs': 1.311177} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.830996] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.831622] env[61986]: INFO nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Took 10.68 seconds to spawn the instance on the hypervisor. [ 823.831834] env[61986]: DEBUG nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.833207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd75045e-b8c0-4597-89fd-24963ba913af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.853486] env[61986]: INFO nova.scheduler.client.report [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Deleted allocations for instance 6c62a1e9-d60c-4720-a554-951bea4b2e18 [ 823.967638] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159678, 'name': PowerOffVM_Task, 'duration_secs': 0.192997} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.967936] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 823.968118] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 824.001836] env[61986]: DEBUG nova.compute.manager [req-66bb9218-f799-47fc-9386-5b2186f8c879 req-ad0c5635-78fc-46ea-b25e-1894de7fe243 service nova] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Received event network-vif-deleted-f1530791-8284-4c8e-a87a-44b3e6145516 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.079103] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.159246] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ab2568-2fa9-5e6b-9152-5ca174dfb26b, 'name': SearchDatastore_Task, 'duration_secs': 0.011769} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.159639] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.160025] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 86c9e45c-b3f1-4004-9acc-190b11a4a926/86c9e45c-b3f1-4004-9acc-190b11a4a926.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 824.160435] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a40b7591-1beb-4938-9f53-a34c8bf62c3d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.170380] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 824.170380] env[61986]: value = "task-1159682" [ 824.170380] env[61986]: _type = "Task" [ 824.170380] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.180712] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.296995] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159675, 'name': CloneVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.306248] env[61986]: DEBUG oslo_vmware.api [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377904} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.306522] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.306722] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 824.306896] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 824.307093] env[61986]: INFO nova.compute.manager [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 824.307700] env[61986]: DEBUG oslo.service.loopingcall [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.307700] env[61986]: DEBUG nova.compute.manager [-] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.307960] env[61986]: DEBUG nova.network.neutron [-] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 824.355378] env[61986]: INFO nova.compute.manager [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Took 30.88 seconds to build instance. [ 824.356188] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.356406] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.356598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.356894] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.356989] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.361382] env[61986]: INFO nova.compute.manager [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Terminating instance [ 824.369343] env[61986]: DEBUG nova.compute.manager [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 824.369343] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 824.370236] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f654a3e-d032-4fdb-88dc-c51389f99ffe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.379935] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.382291] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f2945bf-4797-45ae-b499-d67c99d696a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.386288] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2af328a0-b258-4554-87a7-a7fdb552a9a7 tempest-ServerTagsTestJSON-294943603 tempest-ServerTagsTestJSON-294943603-project-member] Lock "6c62a1e9-d60c-4720-a554-951bea4b2e18" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.974s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.390892] env[61986]: DEBUG oslo_vmware.api [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 824.390892] env[61986]: value = "task-1159683" [ 824.390892] env[61986]: _type = "Task" [ 824.390892] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.403197] env[61986]: DEBUG oslo_vmware.api [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.475910] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.475910] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.476151] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.476340] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.476435] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.476576] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.476790] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.477250] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.477250] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.477487] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.477546] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.483708] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-639aa152-84bd-4082-ab4e-50b9fd04b0ff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.503289] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 824.503289] env[61986]: value = "task-1159684" [ 824.503289] env[61986]: _type = "Task" [ 824.503289] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.513947] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.583582] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.682344] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159682, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.800254] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159675, 'name': CloneVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.861553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41c8fcc8-7758-4571-8fef-3ce0ef534b9e tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.503s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.905101] env[61986]: DEBUG oslo_vmware.api [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159683, 'name': PowerOffVM_Task, 'duration_secs': 0.232344} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.905560] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 824.905796] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 824.905974] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb566bec-b985-4fce-aa09-eef134faf9d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.017460] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159684, 'name': ReconfigVM_Task, 'duration_secs': 0.19587} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.020401] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 825.087830] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.105189] env[61986]: DEBUG nova.network.neutron [-] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.121579] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.123694] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.123694] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Deleting the datastore file [datastore2] ce09db77-1eac-4228-a4f8-228a3b7e69d9 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.123694] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fcf4cf8-76e4-4f19-ae9b-c603baf7229d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.138028] env[61986]: DEBUG oslo_vmware.api [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for the task: (returnval){ [ 825.138028] env[61986]: value = "task-1159686" [ 825.138028] env[61986]: _type = "Task" [ 825.138028] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.147126] env[61986]: DEBUG oslo_vmware.api [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.184411] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702759} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.184675] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 86c9e45c-b3f1-4004-9acc-190b11a4a926/86c9e45c-b3f1-4004-9acc-190b11a4a926.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 825.184910] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.185179] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5360eac3-4cd8-4457-99c5-65d1df5625e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.196369] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 825.196369] env[61986]: value = "task-1159687" [ 825.196369] env[61986]: _type = "Task" [ 825.196369] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.207267] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.298859] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159675, 'name': CloneVM_Task, 'duration_secs': 2.350619} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.301751] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Created linked-clone VM from snapshot [ 825.302758] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fa498a-7e2a-4916-a78d-bbab90b88464 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.311261] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Uploading image 486535fc-ed79-4f4f-a6fb-1cb442ccd708 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 825.336144] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92df07e6-0776-4f09-8f9a-580dbccff393 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.341179] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 825.341179] env[61986]: value = "vm-252330" [ 825.341179] env[61986]: _type = "VirtualMachine" [ 825.341179] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 825.341449] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-abc53453-e875-4c3f-b055-4bf13bf607c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.348203] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763533ad-5716-4ef4-abd0-9e224dcbed0d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.352712] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lease: (returnval){ [ 825.352712] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b768f3-5c96-4e5c-710d-240b3aa46dad" [ 825.352712] env[61986]: _type = "HttpNfcLease" [ 825.352712] env[61986]: } obtained for exporting VM: (result){ [ 825.352712] env[61986]: value = "vm-252330" [ 825.352712] env[61986]: _type = "VirtualMachine" [ 825.352712] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 825.353035] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the lease: (returnval){ [ 825.353035] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b768f3-5c96-4e5c-710d-240b3aa46dad" [ 825.353035] env[61986]: _type = "HttpNfcLease" [ 825.353035] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 825.381914] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.385753] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cddc581-c7a2-49d4-a6f6-ed64b13258d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.390147] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 825.390147] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b768f3-5c96-4e5c-710d-240b3aa46dad" [ 825.390147] env[61986]: _type = "HttpNfcLease" [ 825.390147] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 825.390792] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 825.390792] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b768f3-5c96-4e5c-710d-240b3aa46dad" [ 825.390792] env[61986]: _type = "HttpNfcLease" [ 825.390792] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 825.391515] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a3d76d-2c54-456a-b4c5-68877a418f6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.397848] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d27ad9-8e0e-4004-9b21-c8728f19e938 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.410903] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c12dba-976a-c9d9-9145-9b09142d31f8/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 825.411162] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c12dba-976a-c9d9-9145-9b09142d31f8/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 825.476952] env[61986]: DEBUG nova.compute.provider_tree [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.501710] env[61986]: DEBUG nova.compute.manager [req-a5a1fb67-d8a1-4b0c-a2c3-fe1e9999002a req-833987e0-6993-4153-8c74-cd8acb561572 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Received event network-vif-plugged-e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.501922] env[61986]: DEBUG oslo_concurrency.lockutils [req-a5a1fb67-d8a1-4b0c-a2c3-fe1e9999002a req-833987e0-6993-4153-8c74-cd8acb561572 service nova] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.502137] env[61986]: DEBUG oslo_concurrency.lockutils [req-a5a1fb67-d8a1-4b0c-a2c3-fe1e9999002a req-833987e0-6993-4153-8c74-cd8acb561572 service nova] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.502302] env[61986]: DEBUG oslo_concurrency.lockutils [req-a5a1fb67-d8a1-4b0c-a2c3-fe1e9999002a req-833987e0-6993-4153-8c74-cd8acb561572 service nova] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.502464] env[61986]: DEBUG nova.compute.manager [req-a5a1fb67-d8a1-4b0c-a2c3-fe1e9999002a req-833987e0-6993-4153-8c74-cd8acb561572 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] No waiting events found dispatching network-vif-plugged-e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.502601] env[61986]: WARNING nova.compute.manager [req-a5a1fb67-d8a1-4b0c-a2c3-fe1e9999002a req-833987e0-6993-4153-8c74-cd8acb561572 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Received unexpected event network-vif-plugged-e570192b-e139-47dd-a44e-8f34242b8649 for instance with vm_state building and task_state spawning. [ 825.527805] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:40:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='754ce345-2135-4db7-a907-0c2bf294269e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2075393374',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.528093] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.528298] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.528826] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.528826] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.528950] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.529093] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.529255] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.529764] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.529764] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.529888] env[61986]: DEBUG nova.virt.hardware [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.535822] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfiguring VM instance instance-00000029 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 825.538747] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-630d7d9c-a608-436e-9fa9-5f0084d97fff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.552484] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2a07c55b-e4a2-4439-b6cb-b9c2c188c258 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.561511] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 825.561511] env[61986]: value = "task-1159689" [ 825.561511] env[61986]: _type = "Task" [ 825.561511] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.573778] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159689, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.584924] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.608763] env[61986]: INFO nova.compute.manager [-] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Took 1.30 seconds to deallocate network for instance. [ 825.645368] env[61986]: DEBUG oslo_vmware.api [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Task: {'id': task-1159686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.504067} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.646149] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.646149] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 825.646297] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 825.646396] env[61986]: INFO nova.compute.manager [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Took 1.28 seconds to destroy the instance on the hypervisor. [ 825.646656] env[61986]: DEBUG oslo.service.loopingcall [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.646884] env[61986]: DEBUG nova.compute.manager [-] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.646978] env[61986]: DEBUG nova.network.neutron [-] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 825.707146] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093888} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.707429] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.708211] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d64030b-d825-45d7-9a92-f4d2e1dc944b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.735022] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 86c9e45c-b3f1-4004-9acc-190b11a4a926/86c9e45c-b3f1-4004-9acc-190b11a4a926.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.735022] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf932ac6-d0ca-44dc-8737-dd8997f95639 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.754995] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 825.754995] env[61986]: value = "task-1159690" [ 825.754995] env[61986]: _type = "Task" [ 825.754995] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.762868] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159690, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.777444] env[61986]: DEBUG nova.network.neutron [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Successfully updated port: e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.907111] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.981349] env[61986]: DEBUG nova.scheduler.client.report [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.003666] env[61986]: DEBUG nova.compute.manager [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Received event network-changed-e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.003939] env[61986]: DEBUG nova.compute.manager [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Refreshing instance network info cache due to event network-changed-e570192b-e139-47dd-a44e-8f34242b8649. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 826.005204] env[61986]: DEBUG oslo_concurrency.lockutils [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] Acquiring lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.005204] env[61986]: DEBUG oslo_concurrency.lockutils [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] Acquired lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.005204] env[61986]: DEBUG nova.network.neutron [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Refreshing network info cache for port e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 826.072923] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159689, 'name': ReconfigVM_Task, 'duration_secs': 0.166342} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.073660] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfigured VM instance instance-00000029 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 826.074606] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa66c6bd-c8a8-4a5c-931f-9c27c6ecccda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.104152] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.109362] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aab86eb3-b80b-471e-9435-95c62315e83f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.125557] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.126984] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.128702] env[61986]: DEBUG nova.compute.manager [req-e6141627-4cf6-4485-9151-82b59d398d90 req-a12a19b1-f5f6-4f3c-99f4-a6c50edacd87 service nova] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Received event network-vif-deleted-0495bb14-032d-43ea-9462-1d1d0d326723 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.134738] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 826.134738] env[61986]: value = "task-1159691" [ 826.134738] env[61986]: _type = "Task" [ 826.134738] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.145420] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.271803] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159690, 'name': ReconfigVM_Task, 'duration_secs': 0.283701} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.274032] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 86c9e45c-b3f1-4004-9acc-190b11a4a926/86c9e45c-b3f1-4004-9acc-190b11a4a926.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.274032] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-807f5619-60a2-43a0-867a-cffc206d54d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.283994] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.285806] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 826.285806] env[61986]: value = "task-1159692" [ 826.285806] env[61986]: _type = "Task" [ 826.285806] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.309116] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159692, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.492033] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.492033] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.495431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.328s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.500029] env[61986]: INFO nova.compute.claims [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.560320] env[61986]: DEBUG nova.network.neutron [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.585725] env[61986]: DEBUG oslo_vmware.api [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159679, 'name': PowerOnVM_Task, 'duration_secs': 2.794648} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.585725] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 826.585725] env[61986]: INFO nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Took 10.75 seconds to spawn the instance on the hypervisor. [ 826.585725] env[61986]: DEBUG nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.585725] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f1683e-3bba-4edb-9b49-801a26a2daf8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.652270] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159691, 'name': ReconfigVM_Task, 'duration_secs': 0.323226} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.652750] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.652750] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 50 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 826.664145] env[61986]: DEBUG nova.network.neutron [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.796296] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159692, 'name': Rename_Task, 'duration_secs': 0.179859} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.796636] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.797048] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1517e24-4c4f-404b-921f-e4f69436c320 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.804076] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 826.804076] env[61986]: value = "task-1159693" [ 826.804076] env[61986]: _type = "Task" [ 826.804076] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.812078] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.859568] env[61986]: DEBUG nova.network.neutron [-] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.008866] env[61986]: DEBUG nova.compute.utils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.009992] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.010178] env[61986]: DEBUG nova.network.neutron [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.058852] env[61986]: DEBUG nova.policy [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3481412ab54a0fb386a6d044a9ae7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c88c612fe00a4ad589db67f3016dace6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 827.106496] env[61986]: INFO nova.compute.manager [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Took 27.46 seconds to build instance. [ 827.163284] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442d9f42-c094-4c7c-accc-d555f967d273 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.185851] env[61986]: DEBUG oslo_concurrency.lockutils [req-bd4d51fe-9d6a-46a9-acd8-63a78a446ef0 req-a8c46c91-de96-49f1-ad47-bbb640af402c service nova] Releasing lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.187315] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.187402] env[61986]: DEBUG nova.network.neutron [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 827.189717] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d2ddc7-99dd-4299-9128-b44f15a2d428 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.212632] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 67 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 827.314756] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159693, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.366774] env[61986]: INFO nova.compute.manager [-] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Took 1.72 seconds to deallocate network for instance. [ 827.428023] env[61986]: INFO nova.compute.manager [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Rebuilding instance [ 827.445782] env[61986]: DEBUG nova.network.neutron [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Successfully created port: 8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.489724] env[61986]: DEBUG nova.compute.manager [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.490104] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d70294a-bf8e-4418-9795-0fd94cfbb42f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.517050] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.609632] env[61986]: DEBUG oslo_concurrency.lockutils [None req-06c24a17-190b-40f7-93a4-1e092cc04a1b tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.616s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.772729] env[61986]: DEBUG nova.network.neutron [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Port dd125b45-dffb-4a25-af78-4a9326408be0 binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 827.818525] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159693, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.832026] env[61986]: DEBUG nova.network.neutron [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.878245] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.965601] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.966474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.967105] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.967397] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.968128] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.971709] env[61986]: INFO nova.compute.manager [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Terminating instance [ 827.979148] env[61986]: DEBUG nova.compute.manager [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.979148] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 827.980780] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7e7c1c-3103-4808-8018-6aa4301ac86e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.987970] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 827.988304] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed4f75da-b79d-4de5-a151-34d1c6b8dedc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.991280] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36eb0876-2d58-4a59-ab33-0b4bc58c0d2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.001933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075e8752-4fd2-4c7c-9c82-9be17ad4f596 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.003902] env[61986]: DEBUG oslo_vmware.api [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 828.003902] env[61986]: value = "task-1159694" [ 828.003902] env[61986]: _type = "Task" [ 828.003902] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.005445] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 828.005804] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed16fd75-cb30-440d-9dd1-e27693043377 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.047661] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e97d84-3d70-4ced-8d22-dffce86b177e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.057595] env[61986]: DEBUG oslo_vmware.api [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.057595] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 828.057595] env[61986]: value = "task-1159695" [ 828.057595] env[61986]: _type = "Task" [ 828.057595] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.066748] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e81ad7-0d6e-4ba8-85ee-6d33ae0558b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.073895] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.083887] env[61986]: DEBUG nova.compute.provider_tree [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.112571] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.167467] env[61986]: DEBUG nova.network.neutron [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updating instance_info_cache with network_info: [{"id": "e570192b-e139-47dd-a44e-8f34242b8649", "address": "fa:16:3e:f5:c3:f3", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape570192b-e1", "ovs_interfaceid": "e570192b-e139-47dd-a44e-8f34242b8649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.321209] env[61986]: DEBUG oslo_vmware.api [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159693, 'name': PowerOnVM_Task, 'duration_secs': 1.082584} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.321414] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.321629] env[61986]: INFO nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Took 9.92 seconds to spawn the instance on the hypervisor. [ 828.321831] env[61986]: DEBUG nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.322715] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78986156-e660-4c6b-b423-652b2ac36abe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.515175] env[61986]: DEBUG oslo_vmware.api [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159694, 'name': PowerOffVM_Task, 'duration_secs': 0.228017} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.517160] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 828.517516] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 828.519560] env[61986]: DEBUG nova.compute.manager [req-d3454f2a-5b95-4c90-b75c-b550a4abaa8c req-82574b4c-4360-47b3-9552-10a34e82a68d service nova] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Received event network-vif-deleted-9ffa532a-af81-442f-9253-1c78d86bbd6f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.519997] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70c685fd-20bf-4b42-ac27-93fb459cd34e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.548180] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.567989] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159695, 'name': PowerOffVM_Task, 'duration_secs': 0.198477} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.570320] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 828.570813] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.573391] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e1661b-f2c2-4d40-8f32-e59eb21d13df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.579078] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.579997] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.579997] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.579997] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.579997] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.583026] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.583026] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.583026] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.583026] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.583026] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.583354] env[61986]: DEBUG nova.virt.hardware [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.583354] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c191989d-eb7e-4b01-91ed-dd4a0c482fed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.588209] env[61986]: DEBUG nova.scheduler.client.report [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.592211] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 828.593606] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c9bd849-c4b5-41ff-94ce-b733ed928d39 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.599680] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab639d96-7444-4fb9-9088-baa199f728f0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.604446] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 828.604659] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 828.604874] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Deleting the datastore file [datastore1] ebf5fee5-7f1b-4537-aec3-77a8a963670b {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.605459] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae5c52cb-ce37-4749-bc92-7e5c4a1d67d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.619459] env[61986]: DEBUG oslo_vmware.api [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for the task: (returnval){ [ 828.619459] env[61986]: value = "task-1159698" [ 828.619459] env[61986]: _type = "Task" [ 828.619459] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.630025] env[61986]: DEBUG oslo_vmware.api [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.638352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.660393] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 828.660651] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 828.661023] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleting the datastore file [datastore1] 6f20dd8e-9289-458d-bed6-f2ef9daaa917 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.661428] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cae62ea-3d7c-4938-b7bf-d01ce2b2c145 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.667965] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 828.667965] env[61986]: value = "task-1159699" [ 828.667965] env[61986]: _type = "Task" [ 828.667965] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.671497] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.671802] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Instance network_info: |[{"id": "e570192b-e139-47dd-a44e-8f34242b8649", "address": "fa:16:3e:f5:c3:f3", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape570192b-e1", "ovs_interfaceid": "e570192b-e139-47dd-a44e-8f34242b8649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.672207] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:c3:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e570192b-e139-47dd-a44e-8f34242b8649', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.679636] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Creating folder: Project (c0db18c866dc4ffb8bcb050f8ec6021d). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.680408] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-472378cf-f8c8-4da6-ad13-c9194fa188cc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.685766] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159699, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.694725] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Created folder: Project (c0db18c866dc4ffb8bcb050f8ec6021d) in parent group-v252271. [ 828.694942] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Creating folder: Instances. Parent ref: group-v252331. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.695219] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46f0c34a-792b-49fb-ae8c-131d9d6131a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.704198] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Created folder: Instances in parent group-v252331. [ 828.704562] env[61986]: DEBUG oslo.service.loopingcall [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.704841] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.705076] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba17a840-5024-4601-b110-ed34732b386d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.725204] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.725204] env[61986]: value = "task-1159702" [ 828.725204] env[61986]: _type = "Task" [ 828.725204] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.732684] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159702, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.800025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.800025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.800025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.846730] env[61986]: INFO nova.compute.manager [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Took 27.45 seconds to build instance. [ 829.095464] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.095464] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.098138] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.110s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.098787] env[61986]: DEBUG nova.objects.instance [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lazy-loading 'resources' on Instance uuid 4bb588b3-c5e1-4619-b77b-6734d6516170 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.135249] env[61986]: DEBUG oslo_vmware.api [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Task: {'id': task-1159698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166937} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.135713] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.135790] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 829.136016] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 829.136300] env[61986]: INFO nova.compute.manager [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 829.136589] env[61986]: DEBUG oslo.service.loopingcall [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.136832] env[61986]: DEBUG nova.compute.manager [-] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.137095] env[61986]: DEBUG nova.network.neutron [-] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 829.178623] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159699, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157853} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.180268] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.180475] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 829.180657] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 829.192509] env[61986]: DEBUG nova.network.neutron [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Successfully updated port: 8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.236353] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159702, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.351237] env[61986]: DEBUG oslo_concurrency.lockutils [None req-43bbea10-3249-455b-95fb-0a3c07bd3827 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.018s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.609037] env[61986]: DEBUG nova.compute.utils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.617835] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.617998] env[61986]: DEBUG nova.network.neutron [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.674893] env[61986]: DEBUG nova.policy [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5afe2dd2a41143a4a204e889278df0e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c94c168de68c4608b772cf6fa6012dcb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 829.696179] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.696780] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.696780] env[61986]: DEBUG nova.network.neutron [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.742828] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159702, 'name': CreateVM_Task, 'duration_secs': 0.532597} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.743867] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.743917] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.744272] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.744849] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.745704] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d4d00fa-c6e8-40dd-a3a5-33252fe2e771 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.752760] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 829.752760] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ab08bc-5fb5-5aa0-7ba5-384e1ef30a93" [ 829.752760] env[61986]: _type = "Task" [ 829.752760] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.762739] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ab08bc-5fb5-5aa0-7ba5-384e1ef30a93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.852867] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.894629] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.894909] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.895508] env[61986]: DEBUG nova.network.neutron [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.005897] env[61986]: DEBUG nova.network.neutron [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Successfully created port: 78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.112594] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.139297] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb93091-3079-43b9-bc1e-e18fc64348ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.151377] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055ad295-fdee-48c5-9ce2-3acf9ddd2c71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.189112] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a53a9f-f993-4cfc-9e4b-00e491b8d597 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.199344] env[61986]: DEBUG nova.network.neutron [-] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.202256] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8646fa0-6e63-4171-b08c-dd82f023df8b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.222702] env[61986]: DEBUG nova.compute.provider_tree [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.234637] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.235113] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.235226] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.235544] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.235749] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.235856] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.236090] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.236282] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.236417] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.236585] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.236759] env[61986]: DEBUG nova.virt.hardware [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.238252] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9f1116-80b9-41c5-b04f-0e72a97e9f34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.249738] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c253f37-c6bb-4090-9d5a-04c457ff6297 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.266552] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:5a:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd269a844-d4bf-4b1c-bde6-bb3011fa2d06', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.274270] env[61986]: DEBUG oslo.service.loopingcall [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.279122] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.279401] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ab08bc-5fb5-5aa0-7ba5-384e1ef30a93, 'name': SearchDatastore_Task, 'duration_secs': 0.010191} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.279609] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d3cfb6c-146a-427d-a7ec-59efef31cbf4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.294182] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.294271] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.294616] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.294764] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.294940] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.295968] env[61986]: DEBUG nova.network.neutron [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.297819] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8a14196-0dfc-440c-a52a-efa32f9286cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.306105] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.306105] env[61986]: value = "task-1159703" [ 830.306105] env[61986]: _type = "Task" [ 830.306105] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.313972] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.314175] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.315403] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b313f88a-53d7-48a5-b9f6-de17025f2bd1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.321216] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159703, 'name': CreateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.324138] env[61986]: DEBUG nova.compute.manager [req-0dd10d1a-be4e-45ad-998e-1fcbf1c9757a req-df022ec6-d3e4-4e87-b521-af1e7617cb85 service nova] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Received event network-vif-deleted-5ee8af5d-5341-4ee4-9245-cfa7dc703dca {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.325885] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 830.325885] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523fbd71-0506-8a7d-a95a-eb3829e4bef3" [ 830.325885] env[61986]: _type = "Task" [ 830.325885] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.335743] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523fbd71-0506-8a7d-a95a-eb3829e4bef3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.387799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.669813] env[61986]: DEBUG nova.network.neutron [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.708318] env[61986]: INFO nova.compute.manager [-] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Took 1.57 seconds to deallocate network for instance. [ 830.709751] env[61986]: DEBUG nova.network.neutron [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.731030] env[61986]: DEBUG nova.scheduler.client.report [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.816500] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159703, 'name': CreateVM_Task, 'duration_secs': 0.372323} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.816713] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.817484] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.817671] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.818061] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.818312] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8ac4536-e72d-41c4-b72e-5005c8556b26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.823524] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 830.823524] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5279f97c-448c-2dee-9693-44a66752804d" [ 830.823524] env[61986]: _type = "Task" [ 830.823524] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.836169] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523fbd71-0506-8a7d-a95a-eb3829e4bef3, 'name': SearchDatastore_Task, 'duration_secs': 0.01146} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.840821] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5279f97c-448c-2dee-9693-44a66752804d, 'name': SearchDatastore_Task, 'duration_secs': 0.009779} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.841204] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d11e9f0-566e-4b4e-bae9-b0585cfa9301 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.845836] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.845836] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.845836] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.848316] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 830.848316] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52282555-cb99-d799-435d-2d885464e62f" [ 830.848316] env[61986]: _type = "Task" [ 830.848316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.859473] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52282555-cb99-d799-435d-2d885464e62f, 'name': SearchDatastore_Task, 'duration_secs': 0.00864} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.860289] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.860289] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 9e57f8f9-6e9e-45fb-91d1-132490e930ae/9e57f8f9-6e9e-45fb-91d1-132490e930ae.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 830.860450] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.860618] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.860848] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65eb89d9-6134-4e3f-bbb0-ccbb30117e53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.862821] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c12b936-6d48-42a6-933b-bef232952b1a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.869952] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 830.869952] env[61986]: value = "task-1159704" [ 830.869952] env[61986]: _type = "Task" [ 830.869952] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.871348] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.871560] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.875235] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfa20ef4-269c-4a32-8109-10bceab873bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.882615] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159704, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.884142] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 830.884142] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5261cead-f779-fe98-3cde-31aa6728d16a" [ 830.884142] env[61986]: _type = "Task" [ 830.884142] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.892881] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5261cead-f779-fe98-3cde-31aa6728d16a, 'name': SearchDatastore_Task, 'duration_secs': 0.008088} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.893747] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69725561-7a92-447f-a3ef-b0b3f36fd18b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.899316] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 830.899316] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bfa5af-0df1-0658-efb0-0a3441e636e5" [ 830.899316] env[61986]: _type = "Task" [ 830.899316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.907608] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bfa5af-0df1-0658-efb0-0a3441e636e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.133021] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.162781] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.163148] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.163534] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.163534] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.163635] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.163744] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.164032] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.164223] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.164405] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.164597] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.164982] env[61986]: DEBUG nova.virt.hardware [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.166112] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb3dd25-c1c6-4e5c-ba35-2789f5738dd3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.175096] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.182038] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea1b347-4d7d-4994-bb56-ce9014817920 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.217582] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.218047] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Instance network_info: |[{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.218567] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:37:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bca3360-2dd5-4113-85c7-5e2e44864b07', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.228318] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating folder: Project (c88c612fe00a4ad589db67f3016dace6). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 831.229652] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.229891] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8445c648-7a99-45b5-beb6-2cf9fb0c7656 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.237530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.139s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.241796] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.658s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.243538] env[61986]: INFO nova.compute.claims [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.246633] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created folder: Project (c88c612fe00a4ad589db67f3016dace6) in parent group-v252271. [ 831.246888] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating folder: Instances. Parent ref: group-v252335. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 831.247820] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acfe0dcd-9bae-4143-bd43-935c3c8a2824 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.259165] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created folder: Instances in parent group-v252335. [ 831.259536] env[61986]: DEBUG oslo.service.loopingcall [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.259694] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 831.261453] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31f8c7e4-b112-4ceb-9dee-63d6cc1eec39 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.284720] env[61986]: INFO nova.scheduler.client.report [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Deleted allocations for instance 4bb588b3-c5e1-4619-b77b-6734d6516170 [ 831.294878] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.294878] env[61986]: value = "task-1159707" [ 831.294878] env[61986]: _type = "Task" [ 831.294878] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.303913] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159707, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.370503] env[61986]: DEBUG nova.compute.manager [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Received event network-vif-plugged-8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.370730] env[61986]: DEBUG oslo_concurrency.lockutils [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.370975] env[61986]: DEBUG oslo_concurrency.lockutils [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.371301] env[61986]: DEBUG oslo_concurrency.lockutils [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.371474] env[61986]: DEBUG nova.compute.manager [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] No waiting events found dispatching network-vif-plugged-8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.371663] env[61986]: WARNING nova.compute.manager [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Received unexpected event network-vif-plugged-8bca3360-2dd5-4113-85c7-5e2e44864b07 for instance with vm_state building and task_state spawning. [ 831.372327] env[61986]: DEBUG nova.compute.manager [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Received event network-changed-8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.372327] env[61986]: DEBUG nova.compute.manager [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Refreshing instance network info cache due to event network-changed-8bca3360-2dd5-4113-85c7-5e2e44864b07. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 831.372327] env[61986]: DEBUG oslo_concurrency.lockutils [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.372517] env[61986]: DEBUG oslo_concurrency.lockutils [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.372701] env[61986]: DEBUG nova.network.neutron [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Refreshing network info cache for port 8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.387552] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159704, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501832} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.387826] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 9e57f8f9-6e9e-45fb-91d1-132490e930ae/9e57f8f9-6e9e-45fb-91d1-132490e930ae.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.388088] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.389075] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af085719-044f-4dc8-b50e-1958f918a968 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.397160] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 831.397160] env[61986]: value = "task-1159708" [ 831.397160] env[61986]: _type = "Task" [ 831.397160] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.412327] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.417768] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bfa5af-0df1-0658-efb0-0a3441e636e5, 'name': SearchDatastore_Task, 'duration_secs': 0.008199} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.418093] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.418376] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.418681] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a0cb4dc-4f2c-482f-bd57-28b5cf31684d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.427369] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 831.427369] env[61986]: value = "task-1159709" [ 831.427369] env[61986]: _type = "Task" [ 831.427369] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.436059] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.715542] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f5a9e2-4384-4706-a4db-ff2faf5b8294 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.738384] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a0f822-2fa9-45d9-90ed-313a3a00912c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.747216] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 83 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 831.778116] env[61986]: DEBUG nova.network.neutron [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Successfully updated port: 78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.798809] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3c1cf7c6-63af-46a0-962b-96f3a0cf03bb tempest-TenantUsagesTestJSON-1111729128 tempest-TenantUsagesTestJSON-1111729128-project-member] Lock "4bb588b3-c5e1-4619-b77b-6734d6516170" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.308s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.623335] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 832.626070] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.626070] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.626276] env[61986]: DEBUG nova.network.neutron [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.627852] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159707, 'name': CreateVM_Task, 'duration_secs': 0.395549} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.636223] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93f45528-c25e-4b56-a19f-03b67f1b13e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.636223] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 832.637952] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.638197] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.638896] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.642292] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad25b903-5eff-4b24-8b88-b9d096c4f66e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.648841] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063792} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.649192] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454544} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.654021] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.654342] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 832.654537] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.654815] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 832.654815] env[61986]: value = "task-1159710" [ 832.654815] env[61986]: _type = "Task" [ 832.654815] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.659472] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd3f6af-01cb-420f-ae4e-0c11358f451e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.661279] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb52deb3-1129-4b47-b6a1-2d2747709c19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.667296] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 832.667296] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5209863e-c303-b800-28d9-4aaeed778c45" [ 832.667296] env[61986]: _type = "Task" [ 832.667296] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.695660] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 9e57f8f9-6e9e-45fb-91d1-132490e930ae/9e57f8f9-6e9e-45fb-91d1-132490e930ae.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.697970] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159710, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.697970] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 832.697970] env[61986]: value = "task-1159711" [ 832.697970] env[61986]: _type = "Task" [ 832.697970] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.700318] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de7a0b45-f8d5-4ac0-902b-c0072f6d4171 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.723400] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5209863e-c303-b800-28d9-4aaeed778c45, 'name': SearchDatastore_Task, 'duration_secs': 0.017375} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.725676] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.728247] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.728577] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.728842] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.729034] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.732104] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4215551f-46a4-4871-b013-5197f5a2af1a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.734194] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 832.734194] env[61986]: value = "task-1159712" [ 832.734194] env[61986]: _type = "Task" [ 832.734194] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.737655] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159711, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.745914] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.745914] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.749490] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5eaec72-5635-4cd2-9d29-36b871c8a13b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.755886] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.761018] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 832.761018] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d883b8-e592-0707-b09d-2f9b115cdfd8" [ 832.761018] env[61986]: _type = "Task" [ 832.761018] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.767741] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d883b8-e592-0707-b09d-2f9b115cdfd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.847340] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c12dba-976a-c9d9-9145-9b09142d31f8/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 832.851489] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a594256b-d9db-4ade-9404-b87792e2da6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.864382] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c12dba-976a-c9d9-9145-9b09142d31f8/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 832.864576] env[61986]: ERROR oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c12dba-976a-c9d9-9145-9b09142d31f8/disk-0.vmdk due to incomplete transfer. [ 832.864816] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a457fe22-525d-49d8-bb32-2164d15e38bd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.875439] env[61986]: DEBUG oslo_vmware.rw_handles [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c12dba-976a-c9d9-9145-9b09142d31f8/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 832.875577] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Uploaded image 486535fc-ed79-4f4f-a6fb-1cb442ccd708 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 832.878070] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 832.878343] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c081ca8f-59c7-46fa-a66b-39fda8a1c53d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.891326] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 832.891326] env[61986]: value = "task-1159713" [ 832.891326] env[61986]: _type = "Task" [ 832.891326] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.910110] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159713, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.942196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.942196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.177367] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159710, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.218486] env[61986]: DEBUG nova.network.neutron [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.236294] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072991} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.237506] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.238629] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48b0fb1-9105-4540-8def-b0b57a5d4424 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.262662] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.287417] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.293199] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36e87b70-ed44-465b-8e9e-f78f680c0b8a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.316365] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d883b8-e592-0707-b09d-2f9b115cdfd8, 'name': SearchDatastore_Task, 'duration_secs': 0.009201} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.320091] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8399bec2-d2cc-477f-ab94-f40dcfc0128d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.323055] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 833.323055] env[61986]: value = "task-1159714" [ 833.323055] env[61986]: _type = "Task" [ 833.323055] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.323277] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-680b3345-578c-4e2b-8def-90b8ecc0c17e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.332849] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74ad0e6-2543-4cc8-be55-33b60f7524e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.341579] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159714, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.341962] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 833.341962] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3cb77-3766-2d06-3bf1-bed9a2575480" [ 833.341962] env[61986]: _type = "Task" [ 833.341962] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.378650] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0dc90ca-814e-4d6a-9c66-6321ced2fda5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.385171] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3cb77-3766-2d06-3bf1-bed9a2575480, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.385790] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.386077] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f/92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.386351] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15a45715-463f-4613-aba4-4b43b4900210 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.391480] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab223a0-fce8-4d46-ba23-f41e70c445b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.396754] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 833.396754] env[61986]: value = "task-1159715" [ 833.396754] env[61986]: _type = "Task" [ 833.396754] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.411748] env[61986]: DEBUG nova.compute.provider_tree [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.417494] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159713, 'name': Destroy_Task, 'duration_secs': 0.298932} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.417776] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Destroyed the VM [ 833.418020] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 833.418362] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-eb4a5072-2a5c-4adc-b83d-89c0e00f2947 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.423260] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.428648] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 833.428648] env[61986]: value = "task-1159716" [ 833.428648] env[61986]: _type = "Task" [ 833.428648] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.437876] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159716, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.494204] env[61986]: DEBUG nova.network.neutron [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updated VIF entry in instance network info cache for port 8bca3360-2dd5-4113-85c7-5e2e44864b07. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.494204] env[61986]: DEBUG nova.network.neutron [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.568534] env[61986]: DEBUG nova.compute.manager [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-vif-plugged-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.568811] env[61986]: DEBUG oslo_concurrency.lockutils [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.569686] env[61986]: DEBUG oslo_concurrency.lockutils [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.569946] env[61986]: DEBUG oslo_concurrency.lockutils [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.570185] env[61986]: DEBUG nova.compute.manager [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] No waiting events found dispatching network-vif-plugged-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.570366] env[61986]: WARNING nova.compute.manager [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received unexpected event network-vif-plugged-78729401-474d-4e4f-8159-aff2e908be0b for instance with vm_state building and task_state spawning. [ 833.570532] env[61986]: DEBUG nova.compute.manager [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-changed-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.570752] env[61986]: DEBUG nova.compute.manager [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Refreshing instance network info cache due to event network-changed-78729401-474d-4e4f-8159-aff2e908be0b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 833.571295] env[61986]: DEBUG oslo_concurrency.lockutils [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.598140] env[61986]: DEBUG nova.network.neutron [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.675981] env[61986]: DEBUG oslo_vmware.api [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159710, 'name': PowerOnVM_Task, 'duration_secs': 0.643472} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.676343] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 833.676481] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5568878-6c8f-4a5e-b18e-7cd90b1b0521 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance '40de5f22-0de2-466a-91ab-dcb6ec586dad' progress to 100 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 833.755102] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159712, 'name': ReconfigVM_Task, 'duration_secs': 0.598851} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.755683] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 9e57f8f9-6e9e-45fb-91d1-132490e930ae/9e57f8f9-6e9e-45fb-91d1-132490e930ae.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.756444] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d6228b7-eae4-4af1-8235-f944478224b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.767993] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 833.767993] env[61986]: value = "task-1159717" [ 833.767993] env[61986]: _type = "Task" [ 833.767993] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.777668] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159717, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.839654] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159714, 'name': ReconfigVM_Task, 'duration_secs': 0.441992} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.840475] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 6f20dd8e-9289-458d-bed6-f2ef9daaa917/6f20dd8e-9289-458d-bed6-f2ef9daaa917.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.841982] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14e5566f-e511-4887-9575-ee5145f4632c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.851250] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 833.851250] env[61986]: value = "task-1159718" [ 833.851250] env[61986]: _type = "Task" [ 833.851250] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.873645] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159718, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.910429] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159715, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.915052] env[61986]: DEBUG nova.scheduler.client.report [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.940686] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159716, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.995788] env[61986]: DEBUG oslo_concurrency.lockutils [req-cc7b8fbc-b469-462e-bd89-48781a204317 req-dead1e04-b130-4872-9cff-ad4c81c51037 service nova] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.101502] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.101885] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance network_info: |[{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.102226] env[61986]: DEBUG oslo_concurrency.lockutils [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.102411] env[61986]: DEBUG nova.network.neutron [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Refreshing network info cache for port 78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.104045] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:fd:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e350f83a-f581-4e10-ac16-0b0f7bfd3d38', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78729401-474d-4e4f-8159-aff2e908be0b', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.114045] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating folder: Project (c94c168de68c4608b772cf6fa6012dcb). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.117782] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12a63d75-c9df-4209-a203-a73870d7d2b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.131217] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Created folder: Project (c94c168de68c4608b772cf6fa6012dcb) in parent group-v252271. [ 834.131434] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating folder: Instances. Parent ref: group-v252338. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.132066] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-caea148b-7110-4d2d-9625-6ac3423e6483 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.142218] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Created folder: Instances in parent group-v252338. [ 834.142502] env[61986]: DEBUG oslo.service.loopingcall [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.142736] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.142922] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-732ab47a-cbbf-4b1e-8ceb-b1e5665d5a45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.167110] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.167110] env[61986]: value = "task-1159721" [ 834.167110] env[61986]: _type = "Task" [ 834.167110] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.178549] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159721, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.289222] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159717, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.362268] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159718, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.417659] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159715, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526226} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.417659] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f/92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.418593] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.418593] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09f2d7e0-c9d8-459c-92dd-2ed4fb490487 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.421872] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.180s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.422919] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.548s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.430596] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 834.430596] env[61986]: value = "task-1159722" [ 834.430596] env[61986]: _type = "Task" [ 834.430596] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.446425] env[61986]: DEBUG oslo_vmware.api [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159716, 'name': RemoveSnapshot_Task, 'duration_secs': 0.829556} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.449674] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 834.450240] env[61986]: INFO nova.compute.manager [None req-eac850ce-20e9-473b-b1d4-b6e2be09cb62 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Took 14.35 seconds to snapshot the instance on the hypervisor. [ 834.454524] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159722, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.565518] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "6f9349b7-886a-4077-8f6d-a9800ab353fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.565739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.606433] env[61986]: DEBUG nova.network.neutron [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updated VIF entry in instance network info cache for port 78729401-474d-4e4f-8159-aff2e908be0b. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.606599] env[61986]: DEBUG nova.network.neutron [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.680027] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159721, 'name': CreateVM_Task, 'duration_secs': 0.358018} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.681319] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 834.682048] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.682232] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.682554] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.682816] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af6381d0-7a28-47b9-9d36-5e869703d547 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.688428] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 834.688428] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520211c2-2c3c-647e-6d10-7d53441f9b79" [ 834.688428] env[61986]: _type = "Task" [ 834.688428] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.700310] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520211c2-2c3c-647e-6d10-7d53441f9b79, 'name': SearchDatastore_Task, 'duration_secs': 0.008517} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.700783] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.700863] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.702390] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.702390] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.702390] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.702390] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0adef9db-f5b9-4331-8a25-7125a73b92d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.712574] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.712755] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 834.713485] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4557557e-ef56-4055-a150-5e92cc4ce5a5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.718809] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 834.718809] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52463244-1e9e-a5b3-fdb3-fc4bb0fa691d" [ 834.718809] env[61986]: _type = "Task" [ 834.718809] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.727879] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52463244-1e9e-a5b3-fdb3-fc4bb0fa691d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.782335] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159717, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.862511] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159718, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.930172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "d891bc8a-c8d0-41f2-acbc-229495025958" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.930172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "d891bc8a-c8d0-41f2-acbc-229495025958" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.968063] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159722, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075645} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.968899] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.970440] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f91b74-3d21-4b79-9f8e-743f2cd42bbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.995764] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f/92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.996623] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0026aa9-3234-4143-9002-868928f75b72 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.017985] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 835.017985] env[61986]: value = "task-1159723" [ 835.017985] env[61986]: _type = "Task" [ 835.017985] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.026979] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159723, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.110147] env[61986]: DEBUG oslo_concurrency.lockutils [req-2425e0cf-e88d-4975-8218-ab8a91eb0b65 req-75f861fd-8390-496d-bdf9-0296a345bd18 service nova] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.232155] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52463244-1e9e-a5b3-fdb3-fc4bb0fa691d, 'name': SearchDatastore_Task, 'duration_secs': 0.008857} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.233752] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8859d7d0-78d1-4828-a60d-7fdbba3f1f0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.242888] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 835.242888] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d39840-c03c-8493-e227-467f8e577a1c" [ 835.242888] env[61986]: _type = "Task" [ 835.242888] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.254960] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d39840-c03c-8493-e227-467f8e577a1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.284870] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159717, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.361849] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159718, 'name': Rename_Task, 'duration_secs': 1.235798} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.362048] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 835.362580] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5febf3e-8b5e-473f-be7b-c7abada77c8c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.368967] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 835.368967] env[61986]: value = "task-1159724" [ 835.368967] env[61986]: _type = "Task" [ 835.368967] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.378361] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159724, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.453182] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "d891bc8a-c8d0-41f2-acbc-229495025958" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.524s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.453765] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.459593] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Applying migration context for instance 40de5f22-0de2-466a-91ab-dcb6ec586dad as it has an incoming, in-progress migration 84ab935e-920f-44a1-8d8e-f5b0438906b8. Migration status is finished {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 835.459593] env[61986]: INFO nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating resource usage from migration 84ab935e-920f-44a1-8d8e-f5b0438906b8 [ 835.484907] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa12919e-3891-4da9-a280-4155137864e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.485075] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.485193] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6e0697d4-d862-497c-9bdd-dd1ef2d4272b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.485325] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 977799b4-2793-4513-9447-483146fc7ac4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.485544] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance c1a6ffac-99bd-492c-99e5-f7c46b352d8e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 835.485706] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ce09db77-1eac-4228-a4f8-228a3b7e69d9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 835.485836] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 8438d2a7-975c-4d1f-8906-f960bfadfe39 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 835.485953] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance eaa61f81-2f9b-4d1c-bab8-4363fb71e936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.486075] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6f20dd8e-9289-458d-bed6-f2ef9daaa917 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.486199] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ebf5fee5-7f1b-4537-aec3-77a8a963670b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 835.486309] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 86c9e45c-b3f1-4004-9acc-190b11a4a926 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.486417] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9e57f8f9-6e9e-45fb-91d1-132490e930ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.486523] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.486629] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ee47cab9-5dfd-48ce-ba70-cb800d735b19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.486732] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 182479b8-f72f-4395-99a4-af0d6f91f7d4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 835.533891] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159723, 'name': ReconfigVM_Task, 'duration_secs': 0.270209} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.533891] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f/92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.534250] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc51f8b4-d2f5-4633-a68e-5527c1d3ea12 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.541835] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 835.541835] env[61986]: value = "task-1159725" [ 835.541835] env[61986]: _type = "Task" [ 835.541835] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.549773] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159725, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.755658] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d39840-c03c-8493-e227-467f8e577a1c, 'name': SearchDatastore_Task, 'duration_secs': 0.024249} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.755658] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.755989] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 835.756072] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ffdd44f-b4f6-42a7-9eaa-c14fc36b1f3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.764278] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 835.764278] env[61986]: value = "task-1159726" [ 835.764278] env[61986]: _type = "Task" [ 835.764278] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.771733] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.783852] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "b833adc9-2967-4b0b-81c2-0b8deac20f69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.784595] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.789747] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159717, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.884550] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159724, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.965030] env[61986]: DEBUG nova.compute.utils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.966755] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.966974] env[61986]: DEBUG nova.network.neutron [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 835.992424] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6f38006d-b1f5-4aeb-9124-1e32378c22b2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.052971] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159725, 'name': Rename_Task, 'duration_secs': 0.138194} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.053374] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.053671] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cfa7210-c2b9-46ff-8094-167e2a3e4d48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.062067] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 836.062067] env[61986]: value = "task-1159727" [ 836.062067] env[61986]: _type = "Task" [ 836.062067] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.076620] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159727, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.097693] env[61986]: DEBUG nova.policy [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5706d33724a4197814ddce8838ef757', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '59d8fad2f5734f718a81ee3960842ea5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 836.277024] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480376} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.281534] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 836.281983] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.282123] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c30a79f2-6852-4d2c-86a2-c0c0e6f85417 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.290613] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159717, 'name': Rename_Task, 'duration_secs': 2.30041} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.291962] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.292098] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 836.292098] env[61986]: value = "task-1159728" [ 836.292098] env[61986]: _type = "Task" [ 836.292098] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.292880] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc3ea467-f385-4199-9092-606b002fa649 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.303016] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 836.303016] env[61986]: value = "task-1159729" [ 836.303016] env[61986]: _type = "Task" [ 836.303016] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.305408] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159728, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.313844] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.384288] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159724, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.474018] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.500286] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 021a27ec-cee0-454d-8daf-e6a82bd9330c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 836.574218] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159727, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.660566] env[61986]: DEBUG nova.network.neutron [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Successfully created port: f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.669059] env[61986]: DEBUG nova.network.neutron [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Port dd125b45-dffb-4a25-af78-4a9326408be0 binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 836.669059] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.669059] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.669283] env[61986]: DEBUG nova.network.neutron [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.803919] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159728, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078367} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.804425] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.805694] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e8424c-f3e6-44f4-8158-a041c77ad256 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.836237] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.841612] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-344ae43d-24fb-4ee6-96be-3cfe856254b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.857720] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159729, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.863646] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 836.863646] env[61986]: value = "task-1159730" [ 836.863646] env[61986]: _type = "Task" [ 836.863646] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.872893] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159730, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.879929] env[61986]: DEBUG oslo_vmware.api [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159724, 'name': PowerOnVM_Task, 'duration_secs': 1.219119} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.880313] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.880593] env[61986]: DEBUG nova.compute.manager [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.881456] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3da71d5-992b-47e7-8e58-5ea37dc2f92d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.004492] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a0fa2cbe-d8c3-462a-9744-aaa36a8d314e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 837.073522] env[61986]: DEBUG oslo_vmware.api [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159727, 'name': PowerOnVM_Task, 'duration_secs': 0.526735} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.073800] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.074012] env[61986]: INFO nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 8.53 seconds to spawn the instance on the hypervisor. [ 837.074296] env[61986]: DEBUG nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.075618] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407fdddf-c038-41c3-8323-f17e8c50c261 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.319785] env[61986]: DEBUG oslo_vmware.api [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159729, 'name': PowerOnVM_Task, 'duration_secs': 0.968507} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.320287] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.320591] env[61986]: INFO nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Took 13.60 seconds to spawn the instance on the hypervisor. [ 837.320892] env[61986]: DEBUG nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.321689] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724c7f5f-9a15-4c85-b9f2-9cabbf36c421 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.376563] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159730, 'name': ReconfigVM_Task, 'duration_secs': 0.404886} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.379993] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Reconfigured VM instance instance-00000036 to attach disk [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.381547] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99f15f16-3b7e-4583-bdbd-26463efe2c4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.394430] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 837.394430] env[61986]: value = "task-1159731" [ 837.394430] env[61986]: _type = "Task" [ 837.394430] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.397556] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.408886] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159731, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.469454] env[61986]: DEBUG nova.network.neutron [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.485799] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.510296] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.510628] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.510841] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.511120] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.511338] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.512026] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.512026] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.512026] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.512243] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.512476] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.513812] env[61986]: DEBUG nova.virt.hardware [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.513812] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 46b18b6c-6e40-45b7-9d3f-6177b08a52a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 837.515235] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f96c8-c9d7-4117-980c-2181ecb997b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.524527] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5c7516-933f-40a4-b6f8-bd5c5c5914f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.593399] env[61986]: INFO nova.compute.manager [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 31.22 seconds to build instance. [ 837.837840] env[61986]: INFO nova.compute.manager [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Took 34.83 seconds to build instance. [ 837.907609] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159731, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.972859] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.020324] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 076c8e91-af70-489f-89d5-f598166fafc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.095202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2ce86a5c-e0bf-4845-a79c-9f7d53422a45 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.312s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.343219] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a12a6ab2-3e9e-4139-b348-6ff138009359 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.683s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.409868] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159731, 'name': Rename_Task, 'duration_secs': 0.660295} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.410289] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 838.410603] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5971900e-f20f-4e62-ad12-5abe963edacc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.417548] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 838.417548] env[61986]: value = "task-1159732" [ 838.417548] env[61986]: _type = "Task" [ 838.417548] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.426463] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159732, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.476309] env[61986]: DEBUG nova.compute.manager [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61986) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 838.476565] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.523171] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 52eda388-05f4-416c-ac9d-bd1c1e31ba9c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 838.598391] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.849674] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.929765] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159732, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.030569] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance b9706e51-0f74-4dbc-aab1-ea640ac78c62 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 839.030740] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Migration 84ab935e-920f-44a1-8d8e-f5b0438906b8 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 839.030867] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 40de5f22-0de2-466a-91ab-dcb6ec586dad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 839.123515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.242179] env[61986]: DEBUG nova.network.neutron [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Successfully updated port: f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.379430] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.394701] env[61986]: DEBUG nova.compute.manager [req-f31e95fb-f544-4e4a-a99e-64fa5c86cbce req-a50c2d17-47a1-47b0-bd1d-6f71839b12d2 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Received event network-vif-plugged-f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.394701] env[61986]: DEBUG oslo_concurrency.lockutils [req-f31e95fb-f544-4e4a-a99e-64fa5c86cbce req-a50c2d17-47a1-47b0-bd1d-6f71839b12d2 service nova] Acquiring lock "182479b8-f72f-4395-99a4-af0d6f91f7d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.394701] env[61986]: DEBUG oslo_concurrency.lockutils [req-f31e95fb-f544-4e4a-a99e-64fa5c86cbce req-a50c2d17-47a1-47b0-bd1d-6f71839b12d2 service nova] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.394701] env[61986]: DEBUG oslo_concurrency.lockutils [req-f31e95fb-f544-4e4a-a99e-64fa5c86cbce req-a50c2d17-47a1-47b0-bd1d-6f71839b12d2 service nova] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.394701] env[61986]: DEBUG nova.compute.manager [req-f31e95fb-f544-4e4a-a99e-64fa5c86cbce req-a50c2d17-47a1-47b0-bd1d-6f71839b12d2 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] No waiting events found dispatching network-vif-plugged-f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.395062] env[61986]: WARNING nova.compute.manager [req-f31e95fb-f544-4e4a-a99e-64fa5c86cbce req-a50c2d17-47a1-47b0-bd1d-6f71839b12d2 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Received unexpected event network-vif-plugged-f5e33fe7-af11-4f3e-9b06-db324cbdf3ed for instance with vm_state building and task_state spawning. [ 839.429745] env[61986]: DEBUG oslo_vmware.api [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159732, 'name': PowerOnVM_Task, 'duration_secs': 0.535769} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.432922] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.432922] env[61986]: INFO nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Took 8.30 seconds to spawn the instance on the hypervisor. [ 839.432922] env[61986]: DEBUG nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.434108] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5baf58d2-82ff-4ed6-82a4-956599dac44f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.537941] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 4f13b629-e2a7-4668-9d77-eb638078e246 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 839.747807] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "refresh_cache-182479b8-f72f-4395-99a4-af0d6f91f7d4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.747807] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquired lock "refresh_cache-182479b8-f72f-4395-99a4-af0d6f91f7d4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.747807] env[61986]: DEBUG nova.network.neutron [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.934629] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.935041] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.935256] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.935515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.935749] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.938185] env[61986]: INFO nova.compute.manager [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Terminating instance [ 839.940251] env[61986]: DEBUG nova.compute.manager [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.940502] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.941386] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fa1b07-7cbf-4219-9c83-33e2603313c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.951801] env[61986]: INFO nova.compute.manager [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Took 31.80 seconds to build instance. [ 839.955728] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.955893] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c97224d1-b8ee-4d0a-8892-988593dc97f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.963316] env[61986]: DEBUG oslo_vmware.api [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 839.963316] env[61986]: value = "task-1159733" [ 839.963316] env[61986]: _type = "Task" [ 839.963316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.974632] env[61986]: DEBUG oslo_vmware.api [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.046873] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 98337bb6-9502-4d4c-af00-028659b246bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.304797] env[61986]: DEBUG nova.network.neutron [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.456991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f688d0f-7f11-47b5-98e9-ebd541a713ab tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.798s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.480400] env[61986]: DEBUG oslo_vmware.api [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159733, 'name': PowerOffVM_Task, 'duration_secs': 0.412884} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.480718] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.480913] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.484906] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35f96424-f807-4674-aa22-9bef9acac3ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.551202] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9b539a94-6c39-46f8-b194-27047245d1f2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.564134] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.564134] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.564134] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleting the datastore file [datastore2] 6f20dd8e-9289-458d-bed6-f2ef9daaa917 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.567019] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-940f037a-9c57-4eb7-b41d-02c0ad657250 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.575587] env[61986]: DEBUG oslo_vmware.api [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 840.575587] env[61986]: value = "task-1159735" [ 840.575587] env[61986]: _type = "Task" [ 840.575587] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.583772] env[61986]: DEBUG oslo_vmware.api [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159735, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.596381] env[61986]: DEBUG nova.network.neutron [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Updating instance_info_cache with network_info: [{"id": "f5e33fe7-af11-4f3e-9b06-db324cbdf3ed", "address": "fa:16:3e:a2:7f:b2", "network": {"id": "2aa72323-50d9-474d-93b5-28cce0bb8397", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-747643390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "59d8fad2f5734f718a81ee3960842ea5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "423047aa-c430-4593-a76c-9982c15c81cf", "external-id": "nsx-vlan-transportzone-262", "segmentation_id": 262, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5e33fe7-af", "ovs_interfaceid": "f5e33fe7-af11-4f3e-9b06-db324cbdf3ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.769974] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.770500] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.960066] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.056357] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance d72cd38b-ef14-467b-bf53-97d9e66534c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.085039] env[61986]: DEBUG oslo_vmware.api [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159735, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.345661} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.085543] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.085543] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 841.085696] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.086657] env[61986]: INFO nova.compute.manager [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Took 1.15 seconds to destroy the instance on the hypervisor. [ 841.086657] env[61986]: DEBUG oslo.service.loopingcall [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.086657] env[61986]: DEBUG nova.compute.manager [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.086657] env[61986]: DEBUG nova.network.neutron [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 841.098606] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Releasing lock "refresh_cache-182479b8-f72f-4395-99a4-af0d6f91f7d4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.098953] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Instance network_info: |[{"id": "f5e33fe7-af11-4f3e-9b06-db324cbdf3ed", "address": "fa:16:3e:a2:7f:b2", "network": {"id": "2aa72323-50d9-474d-93b5-28cce0bb8397", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-747643390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "59d8fad2f5734f718a81ee3960842ea5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "423047aa-c430-4593-a76c-9982c15c81cf", "external-id": "nsx-vlan-transportzone-262", "segmentation_id": 262, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5e33fe7-af", "ovs_interfaceid": "f5e33fe7-af11-4f3e-9b06-db324cbdf3ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.099386] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:7f:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '423047aa-c430-4593-a76c-9982c15c81cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5e33fe7-af11-4f3e-9b06-db324cbdf3ed', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.107387] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Creating folder: Project (59d8fad2f5734f718a81ee3960842ea5). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.107659] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2199bb77-b14c-42bb-9e82-6c7a9e8796b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.119405] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Created folder: Project (59d8fad2f5734f718a81ee3960842ea5) in parent group-v252271. [ 841.119405] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Creating folder: Instances. Parent ref: group-v252341. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.119405] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c03b8f9a-f520-408b-9d15-8e417adfb10b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.127865] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Created folder: Instances in parent group-v252341. [ 841.128112] env[61986]: DEBUG oslo.service.loopingcall [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.128308] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.128516] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6138b52-9c51-4920-9011-45481b761d57 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.147803] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.147803] env[61986]: value = "task-1159738" [ 841.147803] env[61986]: _type = "Task" [ 841.147803] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.155144] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159738, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.415576] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "3512cd72-6666-4810-828e-50230956c4b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.416353] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "3512cd72-6666-4810-828e-50230956c4b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.487906] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Received event network-changed-f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.488094] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Refreshing instance network info cache due to event network-changed-f5e33fe7-af11-4f3e-9b06-db324cbdf3ed. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 841.488463] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Acquiring lock "refresh_cache-182479b8-f72f-4395-99a4-af0d6f91f7d4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.488702] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Acquired lock "refresh_cache-182479b8-f72f-4395-99a4-af0d6f91f7d4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.488921] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Refreshing network info cache for port f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.491160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.559886] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a101dc3c-ca6b-4a72-a9b4-051b077a10fd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.658094] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159738, 'name': CreateVM_Task, 'duration_secs': 0.389806} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.658280] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 841.658989] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.659261] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.659597] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.659868] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-010a736a-a44e-4e12-a703-b2841f9457ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.664248] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 841.664248] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528f65b7-7c39-bb90-e5df-535b3233ecb2" [ 841.664248] env[61986]: _type = "Task" [ 841.664248] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.672388] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528f65b7-7c39-bb90-e5df-535b3233ecb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.860853] env[61986]: DEBUG nova.network.neutron [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.063292] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6f9349b7-886a-4077-8f6d-a9800ab353fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.175192] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528f65b7-7c39-bb90-e5df-535b3233ecb2, 'name': SearchDatastore_Task, 'duration_secs': 0.014749} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.175545] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.175815] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.176075] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.176233] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.176416] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.176691] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b685377-8aa8-4fc4-a724-d3d6b5d16206 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.190277] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.190454] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.191385] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f76363e4-68aa-4096-8185-a5813a57f6b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.197094] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 842.197094] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529cb581-3345-1138-03ed-11b0fa32052b" [ 842.197094] env[61986]: _type = "Task" [ 842.197094] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.212361] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529cb581-3345-1138-03ed-11b0fa32052b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.362221] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Updated VIF entry in instance network info cache for port f5e33fe7-af11-4f3e-9b06-db324cbdf3ed. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.362648] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Updating instance_info_cache with network_info: [{"id": "f5e33fe7-af11-4f3e-9b06-db324cbdf3ed", "address": "fa:16:3e:a2:7f:b2", "network": {"id": "2aa72323-50d9-474d-93b5-28cce0bb8397", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-747643390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "59d8fad2f5734f718a81ee3960842ea5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "423047aa-c430-4593-a76c-9982c15c81cf", "external-id": "nsx-vlan-transportzone-262", "segmentation_id": 262, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5e33fe7-af", "ovs_interfaceid": "f5e33fe7-af11-4f3e-9b06-db324cbdf3ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.364531] env[61986]: INFO nova.compute.manager [-] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Took 1.28 seconds to deallocate network for instance. [ 842.554718] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "59cfee64-ef79-4b8f-a703-c8812551d12f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.554910] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.568539] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance b833adc9-2967-4b0b-81c2-0b8deac20f69 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 842.568932] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 842.569083] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 842.712199] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529cb581-3345-1138-03ed-11b0fa32052b, 'name': SearchDatastore_Task, 'duration_secs': 0.031625} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.715236] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20256877-c01c-4899-a962-0453c375f618 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.720395] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 842.720395] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526d5003-1be9-47e2-64ff-f8a9a5725f9a" [ 842.720395] env[61986]: _type = "Task" [ 842.720395] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.729999] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526d5003-1be9-47e2-64ff-f8a9a5725f9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.865329] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Releasing lock "refresh_cache-182479b8-f72f-4395-99a4-af0d6f91f7d4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.865646] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Received event network-changed-8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.865830] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Refreshing instance network info cache due to event network-changed-8bca3360-2dd5-4113-85c7-5e2e44864b07. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.866060] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.866205] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.866363] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Refreshing network info cache for port 8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 842.871989] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.941474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "6488f68e-7b84-4462-aef5-25d02db504f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.941715] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "6488f68e-7b84-4462-aef5-25d02db504f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.966319] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96740b3e-5406-43f4-945c-115a87b0c143 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.974149] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2ced17-e320-4b45-ad02-cced2a60de11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.004908] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdb7ba1-f2f4-4fb3-8b2a-7bd819334e82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.013078] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9884e03f-8058-42d7-82f0-edbc22ba551c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.026027] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.231045] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526d5003-1be9-47e2-64ff-f8a9a5725f9a, 'name': SearchDatastore_Task, 'duration_secs': 0.048582} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.231045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.231384] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 182479b8-f72f-4395-99a4-af0d6f91f7d4/182479b8-f72f-4395-99a4-af0d6f91f7d4.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.231473] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9296339f-ccb5-4a74-b36f-3dd0bf7e1fc4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.238244] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 843.238244] env[61986]: value = "task-1159739" [ 843.238244] env[61986]: _type = "Task" [ 843.238244] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.245208] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159739, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.528771] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.750995] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159739, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439726} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.750995] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 182479b8-f72f-4395-99a4-af0d6f91f7d4/182479b8-f72f-4395-99a4-af0d6f91f7d4.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 843.750995] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.750995] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-198b9c63-63f9-4340-b1ed-05149f7a9dba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.756671] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 843.756671] env[61986]: value = "task-1159740" [ 843.756671] env[61986]: _type = "Task" [ 843.756671] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.764800] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.898717] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updated VIF entry in instance network info cache for port 8bca3360-2dd5-4113-85c7-5e2e44864b07. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.898717] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.033862] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 844.034193] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.611s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.034530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.089s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.036192] env[61986]: INFO nova.compute.claims [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.216828] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.216828] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Cleaning up deleted instances {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 844.269035] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057552} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.269035] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.269035] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa48ee7-8f02-4893-902b-6ad34feb5b8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.298790] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 182479b8-f72f-4395-99a4-af0d6f91f7d4/182479b8-f72f-4395-99a4-af0d6f91f7d4.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.299233] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f2a0edb-1cec-469f-b57f-f5d8034b8a70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.319872] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 844.319872] env[61986]: value = "task-1159741" [ 844.319872] env[61986]: _type = "Task" [ 844.319872] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.328516] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159741, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.401304] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.401876] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Received event network-changed-e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.401876] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Refreshing instance network info cache due to event network-changed-e570192b-e139-47dd-a44e-8f34242b8649. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.402345] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Acquiring lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.402345] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Acquired lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.402525] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Refreshing network info cache for port e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.724515] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] There are 5 instances to clean {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 844.724882] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 4bb588b3-c5e1-4619-b77b-6734d6516170] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 844.831263] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159741, 'name': ReconfigVM_Task, 'duration_secs': 0.439837} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.831466] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 182479b8-f72f-4395-99a4-af0d6f91f7d4/182479b8-f72f-4395-99a4-af0d6f91f7d4.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.832066] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19223436-b1b5-49bd-953b-8d602b1d6eee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.833703] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-977799b4-2793-4513-9447-483146fc7ac4-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.833918] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.834221] env[61986]: DEBUG nova.objects.instance [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'flavor' on Instance uuid 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.839672] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 844.839672] env[61986]: value = "task-1159742" [ 844.839672] env[61986]: _type = "Task" [ 844.839672] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.847300] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159742, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.217967] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updated VIF entry in instance network info cache for port e570192b-e139-47dd-a44e-8f34242b8649. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 845.218483] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updating instance_info_cache with network_info: [{"id": "e570192b-e139-47dd-a44e-8f34242b8649", "address": "fa:16:3e:f5:c3:f3", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape570192b-e1", "ovs_interfaceid": "e570192b-e139-47dd-a44e-8f34242b8649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.231445] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6c62a1e9-d60c-4720-a554-951bea4b2e18] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 845.338388] env[61986]: DEBUG nova.objects.instance [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'pci_requests' on Instance uuid 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.352089] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159742, 'name': Rename_Task, 'duration_secs': 0.134843} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.353098] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 845.353782] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7795a1eb-60aa-4414-95f1-ee3780cb59ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.360701] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 845.360701] env[61986]: value = "task-1159743" [ 845.360701] env[61986]: _type = "Task" [ 845.360701] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.371396] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159743, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.521780] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a5f567-6008-4c4f-98d0-b0210fa493a1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.529886] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8364b1b-babf-4f97-b9a6-673ca67cd045 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.561028] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99f21ae-ff14-40ab-b38a-ebec21abb341 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.567611] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcd8786-a828-42e6-90a6-e2dc30ea9382 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.583324] env[61986]: DEBUG nova.compute.provider_tree [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.722464] env[61986]: DEBUG oslo_concurrency.lockutils [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] Releasing lock "refresh_cache-9e57f8f9-6e9e-45fb-91d1-132490e930ae" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.722464] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Received event network-vif-deleted-d269a844-d4bf-4b1c-bde6-bb3011fa2d06 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 845.722660] env[61986]: INFO nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Neutron deleted interface d269a844-d4bf-4b1c-bde6-bb3011fa2d06; detaching it from the instance and deleting it from the info cache [ 845.722941] env[61986]: DEBUG nova.network.neutron [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.735660] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 2377b482-52e7-4054-8d5b-8f00587c3371] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 845.840656] env[61986]: DEBUG nova.objects.base [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Object Instance<977799b4-2793-4513-9447-483146fc7ac4> lazy-loaded attributes: flavor,pci_requests {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 845.840895] env[61986]: DEBUG nova.network.neutron [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 845.873608] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159743, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.950624] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f7ae02b-0da0-46d4-aae0-1633d3036c71 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.116s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.086616] env[61986]: DEBUG nova.scheduler.client.report [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.225983] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42839753-a44b-4308-b2c7-d6656da33743 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.235701] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c447251-6160-4b8d-924c-71ea610c84e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.246504] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 76b7db85-734f-40fe-8d58-2fba89bf205f] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 846.266129] env[61986]: DEBUG nova.compute.manager [req-a8a7dac7-20ca-4fa7-be63-921c90333ad1 req-759770b5-3cda-44fb-9eda-c875a13c98d5 service nova] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Detach interface failed, port_id=d269a844-d4bf-4b1c-bde6-bb3011fa2d06, reason: Instance 6f20dd8e-9289-458d-bed6-f2ef9daaa917 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 846.373874] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159743, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.592076] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.592583] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.595721] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.795s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.597212] env[61986]: INFO nova.compute.claims [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.749892] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6b586395-eb84-44f6-84c8-4a3ef3d64254] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 846.878087] env[61986]: DEBUG oslo_vmware.api [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159743, 'name': PowerOnVM_Task, 'duration_secs': 1.214719} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.878496] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 846.878786] env[61986]: INFO nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Took 9.39 seconds to spawn the instance on the hypervisor. [ 846.879072] env[61986]: DEBUG nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.880245] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963fc775-cfd5-45b9-8497-3b6eb3a3a952 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.098766] env[61986]: DEBUG nova.compute.utils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.100511] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.100698] env[61986]: DEBUG nova.network.neutron [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 847.160953] env[61986]: DEBUG nova.policy [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c6dfc9a8e294015812f4926e07593e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '040056883ae342589d380bc37eaec5a3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 847.253109] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.253283] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Cleaning up deleted instances with incomplete migration {{(pid=61986) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 847.399384] env[61986]: INFO nova.compute.manager [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Took 34.85 seconds to build instance. [ 847.536740] env[61986]: DEBUG nova.network.neutron [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Successfully created port: 1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.605853] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.755331] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.901440] env[61986]: DEBUG oslo_concurrency.lockutils [None req-887c42fa-c042-481f-ad07-207c8b142da5 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.481s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.945330] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-977799b4-2793-4513-9447-483146fc7ac4-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.945648] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.945925] env[61986]: DEBUG nova.objects.instance [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'flavor' on Instance uuid 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.049622] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4773e23a-143a-4720-b2bb-d73769067332 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.057482] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd2c8f0-567f-4d99-b8cd-e2987c4291c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.125054] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d65a59b-09b5-4139-9261-36ffaeb4e939 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.125054] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cfc7678-b5e1-48eb-8168-00338a0d1154 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.125054] env[61986]: DEBUG nova.compute.provider_tree [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.404506] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.530109] env[61986]: DEBUG nova.objects.instance [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'pci_requests' on Instance uuid 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.616545] env[61986]: DEBUG nova.scheduler.client.report [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.620445] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.645745] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.646033] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.646201] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.646387] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.646538] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.646688] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.646895] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.647071] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.647248] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.647413] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.647588] env[61986]: DEBUG nova.virt.hardware [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.648751] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25c91fe-6913-430c-a7da-abba81a13f0f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.657119] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104a184b-42f7-4db7-84e9-e24ac13441f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.928235] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.033164] env[61986]: DEBUG nova.objects.base [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Object Instance<977799b4-2793-4513-9447-483146fc7ac4> lazy-loaded attributes: flavor,pci_requests {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 849.033422] env[61986]: DEBUG nova.network.neutron [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.073769] env[61986]: DEBUG nova.policy [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 849.124498] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.125168] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.128166] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.735s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.129976] env[61986]: INFO nova.compute.claims [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.212096] env[61986]: DEBUG nova.network.neutron [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Successfully updated port: 1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.219741] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "182479b8-f72f-4395-99a4-af0d6f91f7d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.220092] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.220294] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "182479b8-f72f-4395-99a4-af0d6f91f7d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.221164] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.221517] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.224388] env[61986]: INFO nova.compute.manager [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Terminating instance [ 849.227217] env[61986]: DEBUG nova.compute.manager [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.227217] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.228122] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89da3fe-cbad-43bd-87f2-732455dd3c3e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.236858] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.237220] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-658460f6-97ae-4b7e-becf-b05f4b20327c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.243973] env[61986]: DEBUG oslo_vmware.api [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 849.243973] env[61986]: value = "task-1159744" [ 849.243973] env[61986]: _type = "Task" [ 849.243973] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.252769] env[61986]: DEBUG oslo_vmware.api [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.256452] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 849.256674] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 849.352474] env[61986]: DEBUG nova.network.neutron [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Successfully created port: b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.484964] env[61986]: DEBUG nova.compute.manager [req-ee519205-88f5-4815-a968-05e1897cac7c req-c4b923a7-e1ab-43e8-af41-6e447f33f109 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Received event network-vif-plugged-1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 849.485235] env[61986]: DEBUG oslo_concurrency.lockutils [req-ee519205-88f5-4815-a968-05e1897cac7c req-c4b923a7-e1ab-43e8-af41-6e447f33f109 service nova] Acquiring lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.485411] env[61986]: DEBUG oslo_concurrency.lockutils [req-ee519205-88f5-4815-a968-05e1897cac7c req-c4b923a7-e1ab-43e8-af41-6e447f33f109 service nova] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.485741] env[61986]: DEBUG oslo_concurrency.lockutils [req-ee519205-88f5-4815-a968-05e1897cac7c req-c4b923a7-e1ab-43e8-af41-6e447f33f109 service nova] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.485799] env[61986]: DEBUG nova.compute.manager [req-ee519205-88f5-4815-a968-05e1897cac7c req-c4b923a7-e1ab-43e8-af41-6e447f33f109 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] No waiting events found dispatching network-vif-plugged-1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 849.485940] env[61986]: WARNING nova.compute.manager [req-ee519205-88f5-4815-a968-05e1897cac7c req-c4b923a7-e1ab-43e8-af41-6e447f33f109 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Received unexpected event network-vif-plugged-1c018a8d-09c1-4372-b4fd-f35561ab799b for instance with vm_state building and task_state spawning. [ 849.635804] env[61986]: DEBUG nova.compute.utils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.643046] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.643046] env[61986]: DEBUG nova.network.neutron [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.703874] env[61986]: DEBUG nova.policy [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c637e468ca794a1bac4da331337a1793', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0250d1bd7f4147a19dcb32797b5d794b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 849.715129] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "refresh_cache-6f38006d-b1f5-4aeb-9124-1e32378c22b2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.715283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquired lock "refresh_cache-6f38006d-b1f5-4aeb-9124-1e32378c22b2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.715431] env[61986]: DEBUG nova.network.neutron [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.753812] env[61986]: DEBUG oslo_vmware.api [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159744, 'name': PowerOffVM_Task, 'duration_secs': 0.169907} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.754101] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.754277] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.754531] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9de0eacf-d320-4b89-a3fd-86dabc0b4fef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.813903] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.814127] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.814311] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Deleting the datastore file [datastore1] 182479b8-f72f-4395-99a4-af0d6f91f7d4 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.814581] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e8080b8-9a1a-4ece-9cd9-41e3879b5059 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.822043] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.822179] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.822313] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa12919e-3891-4da9-a280-4155137864e1] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 849.829036] env[61986]: DEBUG oslo_vmware.api [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for the task: (returnval){ [ 849.829036] env[61986]: value = "task-1159746" [ 849.829036] env[61986]: _type = "Task" [ 849.829036] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.840620] env[61986]: DEBUG oslo_vmware.api [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.074159] env[61986]: DEBUG nova.network.neutron [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Successfully created port: 9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.140706] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.253815] env[61986]: DEBUG nova.network.neutron [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.344950] env[61986]: DEBUG oslo_vmware.api [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Task: {'id': task-1159746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141191} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.344950] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.345251] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.345381] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.345648] env[61986]: INFO nova.compute.manager [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 850.345919] env[61986]: DEBUG oslo.service.loopingcall [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.348314] env[61986]: DEBUG nova.compute.manager [-] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.348418] env[61986]: DEBUG nova.network.neutron [-] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.497545] env[61986]: DEBUG nova.network.neutron [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Updating instance_info_cache with network_info: [{"id": "1c018a8d-09c1-4372-b4fd-f35561ab799b", "address": "fa:16:3e:bb:6f:b9", "network": {"id": "5e1a7676-cfd9-451a-aa4a-72e619ab16df", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1344310152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "040056883ae342589d380bc37eaec5a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c018a8d-09", "ovs_interfaceid": "1c018a8d-09c1-4372-b4fd-f35561ab799b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.637199] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c0bda-047a-4380-b54d-a584c312515c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.646026] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c2db31-1bda-4dec-83d8-07b572f71409 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.681175] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2a1a5f-cf9e-4376-816a-7d029b4c5e22 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.689234] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d929017-d1f2-4566-9f82-02c520977d8c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.703083] env[61986]: DEBUG nova.compute.provider_tree [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.920366] env[61986]: DEBUG nova.network.neutron [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Successfully updated port: b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.003645] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Releasing lock "refresh_cache-6f38006d-b1f5-4aeb-9124-1e32378c22b2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.003782] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Instance network_info: |[{"id": "1c018a8d-09c1-4372-b4fd-f35561ab799b", "address": "fa:16:3e:bb:6f:b9", "network": {"id": "5e1a7676-cfd9-451a-aa4a-72e619ab16df", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1344310152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "040056883ae342589d380bc37eaec5a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c018a8d-09", "ovs_interfaceid": "1c018a8d-09c1-4372-b4fd-f35561ab799b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.004229] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:6f:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c018a8d-09c1-4372-b4fd-f35561ab799b', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.015336] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Creating folder: Project (040056883ae342589d380bc37eaec5a3). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.015971] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-684e968b-1e00-4d5f-b75f-c89a6623a7ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.028096] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Created folder: Project (040056883ae342589d380bc37eaec5a3) in parent group-v252271. [ 851.028307] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Creating folder: Instances. Parent ref: group-v252344. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 851.028553] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-131246d3-ceca-4ee0-8932-a7f42959e8f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.038496] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Created folder: Instances in parent group-v252344. [ 851.038741] env[61986]: DEBUG oslo.service.loopingcall [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.038940] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 851.039170] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e030dbda-66c2-4d3b-9cc4-e31b300f0e7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.069287] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.069287] env[61986]: value = "task-1159749" [ 851.069287] env[61986]: _type = "Task" [ 851.069287] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.075963] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159749, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.100500] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updating instance_info_cache with network_info: [{"id": "038ca95f-90ee-4bd7-afb9-b514378d344e", "address": "fa:16:3e:dc:61:92", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038ca95f-90", "ovs_interfaceid": "038ca95f-90ee-4bd7-afb9-b514378d344e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.155743] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.180710] env[61986]: DEBUG nova.network.neutron [-] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.186525] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.186680] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.187159] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.187159] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.188727] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.188847] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.189675] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.189675] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.189909] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.190058] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.190968] env[61986]: DEBUG nova.virt.hardware [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.191479] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1efa4da-83d8-4b11-8f42-6487853c2d2b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.200859] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7701e17-762b-4318-a11e-272c66c66eb0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.206112] env[61986]: DEBUG nova.scheduler.client.report [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.250819] env[61986]: DEBUG nova.compute.manager [req-f51d8bfb-9e07-45e5-8f20-e7f7556c3e01 req-dc884680-837e-44e8-8a00-5780872c551f service nova] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Received event network-vif-deleted-f5e33fe7-af11-4f3e-9b06-db324cbdf3ed {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.423625] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.423935] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.424220] env[61986]: DEBUG nova.network.neutron [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 851.577615] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159749, 'name': CreateVM_Task, 'duration_secs': 0.269609} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.581021] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.581021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.581021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.581021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.581021] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9b3815e-a8bd-4963-a47d-842e90b02c3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.584672] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 851.584672] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3c3bd-5faf-b8ad-1e04-8a966736f358" [ 851.584672] env[61986]: _type = "Task" [ 851.584672] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.592552] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3c3bd-5faf-b8ad-1e04-8a966736f358, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.603226] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-aa12919e-3891-4da9-a280-4155137864e1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.603545] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 851.603889] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.604238] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.604569] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.604864] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.605315] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.605610] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.605844] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 851.606098] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.684929] env[61986]: INFO nova.compute.manager [-] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Took 1.34 seconds to deallocate network for instance. [ 851.710863] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.711466] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.715698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.639s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.715988] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.718572] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.811s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.720385] env[61986]: INFO nova.compute.claims [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.756783] env[61986]: INFO nova.scheduler.client.report [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Deleted allocations for instance 8438d2a7-975c-4d1f-8906-f960bfadfe39 [ 851.852250] env[61986]: DEBUG nova.network.neutron [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Successfully updated port: 9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.960469] env[61986]: WARNING nova.network.neutron [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] 44bde880-8646-44c4-8c3c-c1faf7d4a9ce already exists in list: networks containing: ['44bde880-8646-44c4-8c3c-c1faf7d4a9ce']. ignoring it [ 852.002781] env[61986]: DEBUG nova.compute.manager [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Received event network-changed-1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.003095] env[61986]: DEBUG nova.compute.manager [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Refreshing instance network info cache due to event network-changed-1c018a8d-09c1-4372-b4fd-f35561ab799b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 852.003466] env[61986]: DEBUG oslo_concurrency.lockutils [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] Acquiring lock "refresh_cache-6f38006d-b1f5-4aeb-9124-1e32378c22b2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.003709] env[61986]: DEBUG oslo_concurrency.lockutils [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] Acquired lock "refresh_cache-6f38006d-b1f5-4aeb-9124-1e32378c22b2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.003931] env[61986]: DEBUG nova.network.neutron [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Refreshing network info cache for port 1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 852.096926] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3c3bd-5faf-b8ad-1e04-8a966736f358, 'name': SearchDatastore_Task, 'duration_secs': 0.011182} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.097506] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.098534] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.099054] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.099282] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.099483] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.099757] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f197585-5c7d-438c-8c52-c444773a753c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.109154] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.110471] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.110667] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 852.111399] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f27dd8f-4928-4ce7-ad24-29fda952263d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.117070] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 852.117070] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a57a37-0e93-0c68-8ae2-468c10312f8d" [ 852.117070] env[61986]: _type = "Task" [ 852.117070] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.127628] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a57a37-0e93-0c68-8ae2-468c10312f8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.192305] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.216884] env[61986]: DEBUG nova.compute.utils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.218420] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.219467] env[61986]: DEBUG nova.network.neutron [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.266712] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a0d7e613-1e5a-48c9-9d68-8f855114b954 tempest-ServerAddressesNegativeTestJSON-163637192 tempest-ServerAddressesNegativeTestJSON-163637192-project-member] Lock "8438d2a7-975c-4d1f-8906-f960bfadfe39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.246s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.271032] env[61986]: DEBUG nova.policy [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a79defe9334d45e396e1bc6f68da7da2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30cffc57b42847d8a13ac2347ea2ade7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 852.351894] env[61986]: DEBUG nova.network.neutron [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b2150478-f9cd-4dba-94fe-205081032d36", "address": "fa:16:3e:15:6a:9c", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2150478-f9", "ovs_interfaceid": "b2150478-f9cd-4dba-94fe-205081032d36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.357384] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.357384] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.357384] env[61986]: DEBUG nova.network.neutron [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.633962] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a57a37-0e93-0c68-8ae2-468c10312f8d, 'name': SearchDatastore_Task, 'duration_secs': 0.028051} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.635200] env[61986]: DEBUG nova.network.neutron [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Successfully created port: 3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.639208] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81f39889-ac78-4598-b43a-c5b5d3c87256 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.649722] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 852.649722] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ac99e1-e9b7-b690-11ca-e052fdc10d72" [ 852.649722] env[61986]: _type = "Task" [ 852.649722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.660328] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ac99e1-e9b7-b690-11ca-e052fdc10d72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.724702] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.830111] env[61986]: DEBUG nova.network.neutron [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Updated VIF entry in instance network info cache for port 1c018a8d-09c1-4372-b4fd-f35561ab799b. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 852.830111] env[61986]: DEBUG nova.network.neutron [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Updating instance_info_cache with network_info: [{"id": "1c018a8d-09c1-4372-b4fd-f35561ab799b", "address": "fa:16:3e:bb:6f:b9", "network": {"id": "5e1a7676-cfd9-451a-aa4a-72e619ab16df", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1344310152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "040056883ae342589d380bc37eaec5a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c018a8d-09", "ovs_interfaceid": "1c018a8d-09c1-4372-b4fd-f35561ab799b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.854258] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.854906] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.855082] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.855994] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e15318-5e54-45d8-848b-a8ac03475542 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.886733] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.886961] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.887209] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.887339] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.887493] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.887645] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.887888] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.888081] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.888275] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.888418] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.888592] env[61986]: DEBUG nova.virt.hardware [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.895827] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfiguring VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 852.898671] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0d8c4c4-8b89-46ac-ab53-02889b1878db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.913487] env[61986]: DEBUG nova.network.neutron [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.923716] env[61986]: DEBUG oslo_vmware.api [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 852.923716] env[61986]: value = "task-1159750" [ 852.923716] env[61986]: _type = "Task" [ 852.923716] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.937520] env[61986]: DEBUG oslo_vmware.api [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159750, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.165476] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ac99e1-e9b7-b690-11ca-e052fdc10d72, 'name': SearchDatastore_Task, 'duration_secs': 0.035174} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.169425] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.169739] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 6f38006d-b1f5-4aeb-9124-1e32378c22b2/6f38006d-b1f5-4aeb-9124-1e32378c22b2.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 853.170375] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a4ae9a9-fd84-4203-86ff-52c2609b9407 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.177195] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 853.177195] env[61986]: value = "task-1159751" [ 853.177195] env[61986]: _type = "Task" [ 853.177195] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.188686] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.305693] env[61986]: DEBUG nova.network.neutron [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Updating instance_info_cache with network_info: [{"id": "9456eeae-0969-4fb6-b614-a699facd11f9", "address": "fa:16:3e:ca:f0:1b", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9456eeae-09", "ovs_interfaceid": "9456eeae-0969-4fb6-b614-a699facd11f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.309480] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3d81db-66a7-4d9f-a340-141e551c910b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.318450] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b485bf-a0d9-4b78-a375-b491f1e97d2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.360932] env[61986]: DEBUG oslo_concurrency.lockutils [req-d3a221af-a182-47bc-978b-63e82d0c7991 req-b9e919d7-5c21-4adc-bc79-9494c59af476 service nova] Releasing lock "refresh_cache-6f38006d-b1f5-4aeb-9124-1e32378c22b2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.362933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8baf0ef4-4d4b-41a2-8d61-a47c31123af4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.375999] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48327bee-a989-4502-bb78-7430ae4932cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.392577] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.438045] env[61986]: DEBUG oslo_vmware.api [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.689287] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485928} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.689570] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 6f38006d-b1f5-4aeb-9124-1e32378c22b2/6f38006d-b1f5-4aeb-9124-1e32378c22b2.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 853.689793] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.690080] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf0263bf-2756-424f-89e9-61079cf65f6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.697657] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 853.697657] env[61986]: value = "task-1159752" [ 853.697657] env[61986]: _type = "Task" [ 853.697657] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.706645] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159752, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.738232] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.770077] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.770352] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.770490] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.770668] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.770812] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.770961] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.771255] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.771434] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.771625] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.771769] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.771945] env[61986]: DEBUG nova.virt.hardware [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.774678] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea95c5f-db27-4703-988d-b3753017be21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.783018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f49bd4-e7ee-4e27-856a-3a20e1f0c851 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.813307] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.813695] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Instance network_info: |[{"id": "9456eeae-0969-4fb6-b614-a699facd11f9", "address": "fa:16:3e:ca:f0:1b", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9456eeae-09", "ovs_interfaceid": "9456eeae-0969-4fb6-b614-a699facd11f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.814139] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:f0:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9456eeae-0969-4fb6-b614-a699facd11f9', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.821589] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Creating folder: Project (0250d1bd7f4147a19dcb32797b5d794b). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.821864] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5878ac94-a271-4776-9e40-3ebb042287d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.832628] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Created folder: Project (0250d1bd7f4147a19dcb32797b5d794b) in parent group-v252271. [ 853.832821] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Creating folder: Instances. Parent ref: group-v252347. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.833069] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-847af6e1-d260-44b4-90c8-2e91e33e1778 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.842528] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Created folder: Instances in parent group-v252347. [ 853.842528] env[61986]: DEBUG oslo.service.loopingcall [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.842698] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.842974] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63387c20-caa4-4699-a919-7dc6c5247e4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.862458] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.862458] env[61986]: value = "task-1159755" [ 853.862458] env[61986]: _type = "Task" [ 853.862458] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.869843] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159755, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.895040] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.916758] env[61986]: DEBUG nova.compute.manager [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-vif-plugged-b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.917103] env[61986]: DEBUG oslo_concurrency.lockutils [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.917238] env[61986]: DEBUG oslo_concurrency.lockutils [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.917467] env[61986]: DEBUG oslo_concurrency.lockutils [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.917639] env[61986]: DEBUG nova.compute.manager [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] No waiting events found dispatching network-vif-plugged-b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.917860] env[61986]: WARNING nova.compute.manager [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received unexpected event network-vif-plugged-b2150478-f9cd-4dba-94fe-205081032d36 for instance with vm_state active and task_state None. [ 853.918057] env[61986]: DEBUG nova.compute.manager [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-changed-b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 853.919407] env[61986]: DEBUG nova.compute.manager [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing instance network info cache due to event network-changed-b2150478-f9cd-4dba-94fe-205081032d36. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 853.919407] env[61986]: DEBUG oslo_concurrency.lockutils [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.919407] env[61986]: DEBUG oslo_concurrency.lockutils [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.919407] env[61986]: DEBUG nova.network.neutron [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing network info cache for port b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.935541] env[61986]: DEBUG oslo_vmware.api [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159750, 'name': ReconfigVM_Task, 'duration_secs': 0.689493} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.936121] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.936344] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfigured VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 854.208226] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159752, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060526} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.208463] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.210268] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03dabf6-f988-4b8a-a9c1-edc0401dfb27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.232379] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 6f38006d-b1f5-4aeb-9124-1e32378c22b2/6f38006d-b1f5-4aeb-9124-1e32378c22b2.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.232688] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ddb1ae7-af0c-4934-88ab-93278df16a84 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.252622] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 854.252622] env[61986]: value = "task-1159756" [ 854.252622] env[61986]: _type = "Task" [ 854.252622] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.263740] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159756, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.371962] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159755, 'name': CreateVM_Task, 'duration_secs': 0.311551} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.372488] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 854.373469] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.373844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.374347] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.374727] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0284e2f-9b69-4aa5-acc2-74eb5516b82a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.379733] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 854.379733] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e1552-2b73-62c3-f5ea-ac823f2a828e" [ 854.379733] env[61986]: _type = "Task" [ 854.379733] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.387683] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e1552-2b73-62c3-f5ea-ac823f2a828e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.400392] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.403017] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.410138] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.277s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.410138] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.412682] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.533s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.412682] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.413158] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.775s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.414872] env[61986]: INFO nova.compute.claims [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.441423] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c51e4163-cb49-4a2a-9767-9314c1d81c43 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.495s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.455607] env[61986]: INFO nova.scheduler.client.report [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Deleted allocations for instance ce09db77-1eac-4228-a4f8-228a3b7e69d9 [ 854.458435] env[61986]: INFO nova.scheduler.client.report [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleted allocations for instance c1a6ffac-99bd-492c-99e5-f7c46b352d8e [ 854.486091] env[61986]: DEBUG nova.network.neutron [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Successfully updated port: 3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.690227] env[61986]: DEBUG nova.network.neutron [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updated VIF entry in instance network info cache for port b2150478-f9cd-4dba-94fe-205081032d36. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 854.690695] env[61986]: DEBUG nova.network.neutron [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b2150478-f9cd-4dba-94fe-205081032d36", "address": "fa:16:3e:15:6a:9c", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2150478-f9", "ovs_interfaceid": "b2150478-f9cd-4dba-94fe-205081032d36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.734105] env[61986]: DEBUG nova.compute.manager [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Received event network-vif-plugged-9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.734105] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] Acquiring lock "021a27ec-cee0-454d-8daf-e6a82bd9330c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.734239] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.734407] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.734574] env[61986]: DEBUG nova.compute.manager [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] No waiting events found dispatching network-vif-plugged-9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.734740] env[61986]: WARNING nova.compute.manager [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Received unexpected event network-vif-plugged-9456eeae-0969-4fb6-b614-a699facd11f9 for instance with vm_state building and task_state spawning. [ 854.734902] env[61986]: DEBUG nova.compute.manager [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Received event network-changed-9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.735073] env[61986]: DEBUG nova.compute.manager [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Refreshing instance network info cache due to event network-changed-9456eeae-0969-4fb6-b614-a699facd11f9. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 854.735421] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] Acquiring lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.735421] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] Acquired lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.735550] env[61986]: DEBUG nova.network.neutron [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Refreshing network info cache for port 9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.763136] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.891369] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e1552-2b73-62c3-f5ea-ac823f2a828e, 'name': SearchDatastore_Task, 'duration_secs': 0.009055} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.891678] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.891905] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.892205] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.892303] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.892462] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.892731] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d834929-7af5-4cad-8a51-eeabd5613cb4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.905879] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.906138] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.907332] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10b42f74-108d-4b03-a2b6-db5dde0fd6f0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.913094] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 854.913094] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da2eb9-7405-9055-2bde-c9c9bd95f9ea" [ 854.913094] env[61986]: _type = "Task" [ 854.913094] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.925283] env[61986]: DEBUG nova.compute.utils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.926737] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da2eb9-7405-9055-2bde-c9c9bd95f9ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.927207] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.927509] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.970132] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea1e767b-8b6c-4260-946a-f3fc5648aefb tempest-FloatingIPsAssociationNegativeTestJSON-1189278625 tempest-FloatingIPsAssociationNegativeTestJSON-1189278625-project-member] Lock "ce09db77-1eac-4228-a4f8-228a3b7e69d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.613s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.977864] env[61986]: DEBUG oslo_concurrency.lockutils [None req-871f184b-b438-4ab1-b616-6ad0f5c90052 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "c1a6ffac-99bd-492c-99e5-f7c46b352d8e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.800s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.989826] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.990375] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquired lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.990477] env[61986]: DEBUG nova.network.neutron [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.016518] env[61986]: DEBUG nova.policy [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21b078b477724e9581eba5760ddbd6bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4a585cde0f849f8929d526014ed84d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 855.193890] env[61986]: DEBUG oslo_concurrency.lockutils [req-400828f5-f227-4662-bd89-66173b70309e req-24881f65-0ee3-4fd3-b0eb-0d233446cd5f service nova] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.270302] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159756, 'name': ReconfigVM_Task, 'duration_secs': 0.733702} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.271274] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 6f38006d-b1f5-4aeb-9124-1e32378c22b2/6f38006d-b1f5-4aeb-9124-1e32378c22b2.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.272550] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e9ed462-2c43-4438-a35e-7b72e13464c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.278499] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 855.278499] env[61986]: value = "task-1159757" [ 855.278499] env[61986]: _type = "Task" [ 855.278499] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.294864] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159757, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.423865] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52da2eb9-7405-9055-2bde-c9c9bd95f9ea, 'name': SearchDatastore_Task, 'duration_secs': 0.027137} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.432556] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b452c096-40a3-47dd-b356-097296bbde20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.435899] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.445116] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 855.445116] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522f4b47-8722-b94f-1a11-8490ce9d358b" [ 855.445116] env[61986]: _type = "Task" [ 855.445116] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.452600] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522f4b47-8722-b94f-1a11-8490ce9d358b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.678704] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Successfully created port: 8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.685704] env[61986]: DEBUG nova.network.neutron [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.796915] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159757, 'name': Rename_Task, 'duration_secs': 0.148909} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.797314] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 855.797716] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9da8ed10-efe8-4404-8906-1f4cd103f25b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.804781] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 855.804781] env[61986]: value = "task-1159758" [ 855.804781] env[61986]: _type = "Task" [ 855.804781] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.820422] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.848442] env[61986]: DEBUG nova.network.neutron [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Updated VIF entry in instance network info cache for port 9456eeae-0969-4fb6-b614-a699facd11f9. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 855.848945] env[61986]: DEBUG nova.network.neutron [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Updating instance_info_cache with network_info: [{"id": "9456eeae-0969-4fb6-b614-a699facd11f9", "address": "fa:16:3e:ca:f0:1b", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9456eeae-09", "ovs_interfaceid": "9456eeae-0969-4fb6-b614-a699facd11f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.964171] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522f4b47-8722-b94f-1a11-8490ce9d358b, 'name': SearchDatastore_Task, 'duration_secs': 0.014801} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.964530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.965179] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/021a27ec-cee0-454d-8daf-e6a82bd9330c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 855.965678] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83636599-70ca-4957-a4f0-f3bf3c677d79 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.982579] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 855.982579] env[61986]: value = "task-1159759" [ 855.982579] env[61986]: _type = "Task" [ 855.982579] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.985660] env[61986]: DEBUG nova.compute.manager [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Received event network-vif-plugged-3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.985660] env[61986]: DEBUG oslo_concurrency.lockutils [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.985660] env[61986]: DEBUG oslo_concurrency.lockutils [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.985660] env[61986]: DEBUG oslo_concurrency.lockutils [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.985660] env[61986]: DEBUG nova.compute.manager [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] No waiting events found dispatching network-vif-plugged-3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.986092] env[61986]: WARNING nova.compute.manager [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Received unexpected event network-vif-plugged-3cf0a017-8fc3-4649-bb5a-922ad76c9140 for instance with vm_state building and task_state spawning. [ 855.986092] env[61986]: DEBUG nova.compute.manager [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Received event network-changed-3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.986092] env[61986]: DEBUG nova.compute.manager [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Refreshing instance network info cache due to event network-changed-3cf0a017-8fc3-4649-bb5a-922ad76c9140. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 855.986317] env[61986]: DEBUG oslo_concurrency.lockutils [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] Acquiring lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.996047] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.042943] env[61986]: DEBUG nova.network.neutron [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating instance_info_cache with network_info: [{"id": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "address": "fa:16:3e:80:d7:ec", "network": {"id": "dde733ac-b7c6-4939-a972-c681f487c6b8", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1722442317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30cffc57b42847d8a13ac2347ea2ade7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cf0a017-8f", "ovs_interfaceid": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.127110] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28993884-e10b-4a72-9dbb-cd8ed0a5572e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.135397] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b9b389-e0ee-4692-8f8e-61f574effe40 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.167471] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c25f84b-400d-46b4-9b7b-ca83d4a7762d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.175526] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2034448-14d8-4ab8-9fa7-79932db55ba7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.189249] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.318169] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159758, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.352560] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ed0016a-50c2-438d-a142-921d474ea3f2 req-b7e81ba5-abc5-4710-8fb0-a1b488a0ef5a service nova] Releasing lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.455037] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.496154] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.496489] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.496691] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.496909] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.497101] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.497293] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.497513] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.497714] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.497959] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.498202] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.498397] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.499511] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f849edd2-f3bf-4450-b018-5accef10691a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.509289] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159759, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.512598] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5931d82d-459e-432f-9d30-eafb6c7e4801 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.545619] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Releasing lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.547378] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Instance network_info: |[{"id": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "address": "fa:16:3e:80:d7:ec", "network": {"id": "dde733ac-b7c6-4939-a972-c681f487c6b8", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1722442317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30cffc57b42847d8a13ac2347ea2ade7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cf0a017-8f", "ovs_interfaceid": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.547378] env[61986]: DEBUG oslo_concurrency.lockutils [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] Acquired lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.547378] env[61986]: DEBUG nova.network.neutron [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Refreshing network info cache for port 3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.548682] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:d7:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bafe8721-91d4-4127-b215-d9e8e27947dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cf0a017-8fc3-4649-bb5a-922ad76c9140', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.557682] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Creating folder: Project (30cffc57b42847d8a13ac2347ea2ade7). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 856.558235] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bf8b34d-022b-437f-8c52-3e226e388cb0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.570696] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Created folder: Project (30cffc57b42847d8a13ac2347ea2ade7) in parent group-v252271. [ 856.570939] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Creating folder: Instances. Parent ref: group-v252350. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 856.571209] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c29896e-cac4-4ffd-8bc7-60ec1b61df57 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.587468] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Created folder: Instances in parent group-v252350. [ 856.587725] env[61986]: DEBUG oslo.service.loopingcall [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.587973] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 856.588206] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7980a9e0-e02f-4b3f-820f-ade7bc30e132 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.610649] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.610649] env[61986]: value = "task-1159762" [ 856.610649] env[61986]: _type = "Task" [ 856.610649] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.618267] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159762, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.695027] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.777218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-977799b4-2793-4513-9447-483146fc7ac4-5a6caae3-e994-43ff-b2b4-5e3f89aea659" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.777571] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-5a6caae3-e994-43ff-b2b4-5e3f89aea659" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.778485] env[61986]: DEBUG nova.objects.instance [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'flavor' on Instance uuid 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.818146] env[61986]: DEBUG oslo_vmware.api [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159758, 'name': PowerOnVM_Task, 'duration_secs': 0.5527} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.818146] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 856.818146] env[61986]: INFO nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Took 8.20 seconds to spawn the instance on the hypervisor. [ 856.818146] env[61986]: DEBUG nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.821114] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc53a30-290a-43f3-8ac3-46c49612ff90 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.997095] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553891} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.997382] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/021a27ec-cee0-454d-8daf-e6a82bd9330c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 856.997900] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.000028] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7731b286-53bb-44ec-ba0f-99a689fe3071 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.006482] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 857.006482] env[61986]: value = "task-1159763" [ 857.006482] env[61986]: _type = "Task" [ 857.006482] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.013600] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.121770] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159762, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.199702] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.787s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.200247] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.202792] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.815s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.204180] env[61986]: INFO nova.compute.claims [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.340757] env[61986]: INFO nova.compute.manager [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Took 42.41 seconds to build instance. [ 857.412327] env[61986]: DEBUG nova.network.neutron [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updated VIF entry in instance network info cache for port 3cf0a017-8fc3-4649-bb5a-922ad76c9140. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.412711] env[61986]: DEBUG nova.network.neutron [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating instance_info_cache with network_info: [{"id": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "address": "fa:16:3e:80:d7:ec", "network": {"id": "dde733ac-b7c6-4939-a972-c681f487c6b8", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1722442317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30cffc57b42847d8a13ac2347ea2ade7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cf0a017-8f", "ovs_interfaceid": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.436125] env[61986]: DEBUG nova.objects.instance [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'pci_requests' on Instance uuid 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.516425] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064002} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.516907] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.517831] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2643a3e2-0f2c-45d8-af55-a44f7f45ce79 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.546636] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/021a27ec-cee0-454d-8daf-e6a82bd9330c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.547267] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9e32e57-55e3-45e3-ad82-66eeced0caac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.571349] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 857.571349] env[61986]: value = "task-1159764" [ 857.571349] env[61986]: _type = "Task" [ 857.571349] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.579713] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159764, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.623807] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159762, 'name': CreateVM_Task, 'duration_secs': 0.699011} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.624134] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 857.625115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.625411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.625818] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.626199] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dceacd54-472a-4c95-95bd-f2a2bd552ce4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.630625] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 857.630625] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5273d6ae-6299-dce8-854d-20d72e1c02f4" [ 857.630625] env[61986]: _type = "Task" [ 857.630625] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.638549] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5273d6ae-6299-dce8-854d-20d72e1c02f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.711804] env[61986]: DEBUG nova.compute.utils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.714746] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.715110] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.759366] env[61986]: DEBUG nova.policy [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21b078b477724e9581eba5760ddbd6bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4a585cde0f849f8929d526014ed84d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 857.843346] env[61986]: DEBUG oslo_concurrency.lockutils [None req-29ec4d78-cd64-4b96-9310-474a72d0f7b2 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.648s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.916064] env[61986]: DEBUG oslo_concurrency.lockutils [req-05b17629-311e-44d7-94f4-a802eed44e2d req-b9e90db8-e64c-4046-aef3-5c72774bcde9 service nova] Releasing lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.937624] env[61986]: DEBUG nova.objects.base [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Object Instance<977799b4-2793-4513-9447-483146fc7ac4> lazy-loaded attributes: flavor,pci_requests {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 857.937882] env[61986]: DEBUG nova.network.neutron [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.062619] env[61986]: DEBUG nova.policy [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 858.081117] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159764, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.144662] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5273d6ae-6299-dce8-854d-20d72e1c02f4, 'name': SearchDatastore_Task, 'duration_secs': 0.046416} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.145684] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.146308] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.146597] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.146756] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.147057] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.147385] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21e79254-7734-4615-9c6d-00c32c516d33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.162452] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.162656] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 858.163431] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1738e53f-3dac-4b71-8985-0fecef923048 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.172342] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 858.172342] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528b13ad-052d-2ffc-f746-a7630b976613" [ 858.172342] env[61986]: _type = "Task" [ 858.172342] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.180615] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528b13ad-052d-2ffc-f746-a7630b976613, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.215338] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.258444] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Successfully created port: 18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.292457] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Successfully updated port: 8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 858.346928] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.405165] env[61986]: DEBUG nova.compute.manager [req-e92bb3db-45c2-4394-a1e3-952fba66262f req-099b0179-1078-4140-b557-6f22dd53b9da service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Received event network-vif-plugged-8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.405165] env[61986]: DEBUG oslo_concurrency.lockutils [req-e92bb3db-45c2-4394-a1e3-952fba66262f req-099b0179-1078-4140-b557-6f22dd53b9da service nova] Acquiring lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.405561] env[61986]: DEBUG oslo_concurrency.lockutils [req-e92bb3db-45c2-4394-a1e3-952fba66262f req-099b0179-1078-4140-b557-6f22dd53b9da service nova] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.405837] env[61986]: DEBUG oslo_concurrency.lockutils [req-e92bb3db-45c2-4394-a1e3-952fba66262f req-099b0179-1078-4140-b557-6f22dd53b9da service nova] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.406136] env[61986]: DEBUG nova.compute.manager [req-e92bb3db-45c2-4394-a1e3-952fba66262f req-099b0179-1078-4140-b557-6f22dd53b9da service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] No waiting events found dispatching network-vif-plugged-8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.406658] env[61986]: WARNING nova.compute.manager [req-e92bb3db-45c2-4394-a1e3-952fba66262f req-099b0179-1078-4140-b557-6f22dd53b9da service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Received unexpected event network-vif-plugged-8b44e4ce-d172-4864-90e3-3c6cf9486bfb for instance with vm_state building and task_state spawning. [ 858.520106] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.521319] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.587178] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159764, 'name': ReconfigVM_Task, 'duration_secs': 0.820939} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.591025] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/021a27ec-cee0-454d-8daf-e6a82bd9330c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.591985] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43ed7617-6c33-47b3-9abd-162dabb696e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.598817] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 858.598817] env[61986]: value = "task-1159765" [ 858.598817] env[61986]: _type = "Task" [ 858.598817] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.607647] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159765, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.684667] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528b13ad-052d-2ffc-f746-a7630b976613, 'name': SearchDatastore_Task, 'duration_secs': 0.009529} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.685652] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d215dfc-6032-4829-961b-9d0c6a9c13c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.692156] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 858.692156] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525e73ad-cbb2-73fc-4ebe-655c7b34f120" [ 858.692156] env[61986]: _type = "Task" [ 858.692156] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.709222] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525e73ad-cbb2-73fc-4ebe-655c7b34f120, 'name': SearchDatastore_Task, 'duration_secs': 0.00983} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.713313] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.714041] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] a0fa2cbe-d8c3-462a-9744-aaa36a8d314e/a0fa2cbe-d8c3-462a-9744-aaa36a8d314e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.714460] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-049cf0d1-056c-4688-8805-fd1e049095d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.731109] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 858.731109] env[61986]: value = "task-1159766" [ 858.731109] env[61986]: _type = "Task" [ 858.731109] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.743216] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.797381] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "refresh_cache-46b18b6c-6e40-45b7-9d3f-6177b08a52a5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.797464] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "refresh_cache-46b18b6c-6e40-45b7-9d3f-6177b08a52a5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.797616] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.830031] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa910f0-bbcb-4263-be15-62577cc19313 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.838011] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1f13d0-08d5-4c42-8fe1-f33fac160217 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.871804] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb518f9e-2e46-4a4d-90cb-f3e75b92c905 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.882642] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16af8bc9-600d-4cd0-b2e1-12a3b3ddd7c1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.897996] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.900097] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.114039] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159765, 'name': Rename_Task, 'duration_secs': 0.222783} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.114602] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 859.115013] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0520a34-036a-4fac-a04d-42003bf1af1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.125676] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 859.125676] env[61986]: value = "task-1159767" [ 859.125676] env[61986]: _type = "Task" [ 859.125676] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.142938] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.235010] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.247094] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159766, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.274132] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.274620] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.274620] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.274761] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.274852] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.275064] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.275424] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.275809] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.276016] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.276329] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.276644] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.277622] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcb3653-be3d-4872-a542-6c69debc3342 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.287182] env[61986]: DEBUG nova.compute.manager [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Received event network-changed-8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.287451] env[61986]: DEBUG nova.compute.manager [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Refreshing instance network info cache due to event network-changed-8b44e4ce-d172-4864-90e3-3c6cf9486bfb. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 859.287765] env[61986]: DEBUG oslo_concurrency.lockutils [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] Acquiring lock "refresh_cache-46b18b6c-6e40-45b7-9d3f-6177b08a52a5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.292656] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047dde61-ee4d-437b-aedc-54dc84d49c32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.366188] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 859.421507] env[61986]: ERROR nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [req-6fb26cc5-521f-4102-8c11-1410cb461d4d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6fb26cc5-521f-4102-8c11-1410cb461d4d"}]} [ 859.442304] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Refreshing inventories for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 859.460321] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updating ProviderTree inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 859.460565] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 859.474553] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Refreshing aggregate associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, aggregates: None {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 859.497593] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Refreshing trait associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 859.640485] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159767, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.758833] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594792} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.759271] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] a0fa2cbe-d8c3-462a-9744-aaa36a8d314e/a0fa2cbe-d8c3-462a-9744-aaa36a8d314e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.760010] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.760728] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a6b78ac-1b99-4584-8d70-c5f545404066 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.772594] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 859.772594] env[61986]: value = "task-1159768" [ 859.772594] env[61986]: _type = "Task" [ 859.772594] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.784473] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.798025] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Updating instance_info_cache with network_info: [{"id": "8b44e4ce-d172-4864-90e3-3c6cf9486bfb", "address": "fa:16:3e:ef:56:b7", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b44e4ce-d1", "ovs_interfaceid": "8b44e4ce-d172-4864-90e3-3c6cf9486bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.019867] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8834d36-11d8-430e-8205-53a38f17ccea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.028401] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d528e1-b536-48f9-88c6-ab5a516864b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.066886] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f6d050-3f85-4abe-baae-22451f5a8f6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.075532] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3aad07-ac4c-4a95-8f8d-7f79aa2cadfc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.091590] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.139643] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159767, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.220250] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Successfully updated port: 18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.282699] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.240043} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.282983] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.284577] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebaf1d3-68cb-40b6-a0ea-808c31a65c45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.311116] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] a0fa2cbe-d8c3-462a-9744-aaa36a8d314e/a0fa2cbe-d8c3-462a-9744-aaa36a8d314e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.314270] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "refresh_cache-46b18b6c-6e40-45b7-9d3f-6177b08a52a5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.314270] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Instance network_info: |[{"id": "8b44e4ce-d172-4864-90e3-3c6cf9486bfb", "address": "fa:16:3e:ef:56:b7", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b44e4ce-d1", "ovs_interfaceid": "8b44e4ce-d172-4864-90e3-3c6cf9486bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.314270] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b5456f9-f447-448c-8ba1-066cbd0054e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.329446] env[61986]: DEBUG oslo_concurrency.lockutils [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] Acquired lock "refresh_cache-46b18b6c-6e40-45b7-9d3f-6177b08a52a5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.329446] env[61986]: DEBUG nova.network.neutron [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Refreshing network info cache for port 8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 860.329791] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:56:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b44e4ce-d172-4864-90e3-3c6cf9486bfb', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.337090] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Creating folder: Project (f4a585cde0f849f8929d526014ed84d1). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 860.341563] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0481051a-e20b-4fed-8c07-54b46503de89 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.347563] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 860.347563] env[61986]: value = "task-1159770" [ 860.347563] env[61986]: _type = "Task" [ 860.347563] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.352557] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Created folder: Project (f4a585cde0f849f8929d526014ed84d1) in parent group-v252271. [ 860.352744] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Creating folder: Instances. Parent ref: group-v252353. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 860.355980] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bd8a9c6-2ea8-4a82-92ac-69b4e856002e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.357457] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.364954] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Created folder: Instances in parent group-v252353. [ 860.365214] env[61986]: DEBUG oslo.service.loopingcall [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.365384] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 860.365598] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e19f3f6-7d03-4268-a3ce-5d59abe9ff44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.385528] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.385528] env[61986]: value = "task-1159772" [ 860.385528] env[61986]: _type = "Task" [ 860.385528] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.392837] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159772, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.485496] env[61986]: DEBUG nova.compute.manager [req-2e33fc8c-cbc7-454a-bb51-07765407dae2 req-fdb845ab-b5a2-490e-9f95-70a5a5c4d231 service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Received event network-vif-plugged-18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.485714] env[61986]: DEBUG oslo_concurrency.lockutils [req-2e33fc8c-cbc7-454a-bb51-07765407dae2 req-fdb845ab-b5a2-490e-9f95-70a5a5c4d231 service nova] Acquiring lock "076c8e91-af70-489f-89d5-f598166fafc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.485956] env[61986]: DEBUG oslo_concurrency.lockutils [req-2e33fc8c-cbc7-454a-bb51-07765407dae2 req-fdb845ab-b5a2-490e-9f95-70a5a5c4d231 service nova] Lock "076c8e91-af70-489f-89d5-f598166fafc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.486154] env[61986]: DEBUG oslo_concurrency.lockutils [req-2e33fc8c-cbc7-454a-bb51-07765407dae2 req-fdb845ab-b5a2-490e-9f95-70a5a5c4d231 service nova] Lock "076c8e91-af70-489f-89d5-f598166fafc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.486324] env[61986]: DEBUG nova.compute.manager [req-2e33fc8c-cbc7-454a-bb51-07765407dae2 req-fdb845ab-b5a2-490e-9f95-70a5a5c4d231 service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] No waiting events found dispatching network-vif-plugged-18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.486576] env[61986]: WARNING nova.compute.manager [req-2e33fc8c-cbc7-454a-bb51-07765407dae2 req-fdb845ab-b5a2-490e-9f95-70a5a5c4d231 service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Received unexpected event network-vif-plugged-18e4cc01-51d6-4a7f-b279-4427cedb88ca for instance with vm_state building and task_state spawning. [ 860.585370] env[61986]: DEBUG nova.network.neutron [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Successfully updated port: 5a6caae3-e994-43ff-b2b4-5e3f89aea659 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.630234] env[61986]: DEBUG nova.scheduler.client.report [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 860.630550] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 84 to 85 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 860.630757] env[61986]: DEBUG nova.compute.provider_tree [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.644262] env[61986]: DEBUG oslo_vmware.api [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159767, 'name': PowerOnVM_Task, 'duration_secs': 1.103099} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.644555] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 860.645624] env[61986]: INFO nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Took 9.49 seconds to spawn the instance on the hypervisor. [ 860.645624] env[61986]: DEBUG nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.645932] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde7db8a-d0fd-4b44-974c-c9c7714d11d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.727953] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "refresh_cache-076c8e91-af70-489f-89d5-f598166fafc6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.727953] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "refresh_cache-076c8e91-af70-489f-89d5-f598166fafc6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.728262] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.818516] env[61986]: DEBUG nova.network.neutron [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Updated VIF entry in instance network info cache for port 8b44e4ce-d172-4864-90e3-3c6cf9486bfb. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.818918] env[61986]: DEBUG nova.network.neutron [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Updating instance_info_cache with network_info: [{"id": "8b44e4ce-d172-4864-90e3-3c6cf9486bfb", "address": "fa:16:3e:ef:56:b7", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b44e4ce-d1", "ovs_interfaceid": "8b44e4ce-d172-4864-90e3-3c6cf9486bfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.861267] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.897890] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159772, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.976665] env[61986]: DEBUG nova.compute.manager [req-7340c55b-d4ff-431b-832a-abdeb88424d4 req-979daee3-2ac7-41d4-a2d2-a97f7829329c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-vif-plugged-5a6caae3-e994-43ff-b2b4-5e3f89aea659 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.976892] env[61986]: DEBUG oslo_concurrency.lockutils [req-7340c55b-d4ff-431b-832a-abdeb88424d4 req-979daee3-2ac7-41d4-a2d2-a97f7829329c service nova] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.979449] env[61986]: DEBUG oslo_concurrency.lockutils [req-7340c55b-d4ff-431b-832a-abdeb88424d4 req-979daee3-2ac7-41d4-a2d2-a97f7829329c service nova] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.979449] env[61986]: DEBUG oslo_concurrency.lockutils [req-7340c55b-d4ff-431b-832a-abdeb88424d4 req-979daee3-2ac7-41d4-a2d2-a97f7829329c service nova] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.979449] env[61986]: DEBUG nova.compute.manager [req-7340c55b-d4ff-431b-832a-abdeb88424d4 req-979daee3-2ac7-41d4-a2d2-a97f7829329c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] No waiting events found dispatching network-vif-plugged-5a6caae3-e994-43ff-b2b4-5e3f89aea659 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.984028] env[61986]: WARNING nova.compute.manager [req-7340c55b-d4ff-431b-832a-abdeb88424d4 req-979daee3-2ac7-41d4-a2d2-a97f7829329c service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received unexpected event network-vif-plugged-5a6caae3-e994-43ff-b2b4-5e3f89aea659 for instance with vm_state active and task_state None. [ 861.087306] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.087644] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.087727] env[61986]: DEBUG nova.network.neutron [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.140306] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.937s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.140988] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.143707] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.914s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.147172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.149851] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.752s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.150345] env[61986]: DEBUG nova.objects.instance [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 861.163402] env[61986]: INFO nova.compute.manager [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Took 42.39 seconds to build instance. [ 861.185450] env[61986]: INFO nova.scheduler.client.report [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Deleted allocations for instance ebf5fee5-7f1b-4537-aec3-77a8a963670b [ 861.268949] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.324030] env[61986]: DEBUG oslo_concurrency.lockutils [req-b88b9aaa-55a0-4d7d-8b5d-e45895d1b9e9 req-dbe69d6f-b0e5-403e-9e8f-6910c25faece service nova] Releasing lock "refresh_cache-46b18b6c-6e40-45b7-9d3f-6177b08a52a5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.365532] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159770, 'name': ReconfigVM_Task, 'duration_secs': 0.934049} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.365871] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfigured VM instance instance-0000003a to attach disk [datastore1] a0fa2cbe-d8c3-462a-9744-aaa36a8d314e/a0fa2cbe-d8c3-462a-9744-aaa36a8d314e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.366523] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b341386-28d2-46a0-ace9-e8fdc0c946f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.372636] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 861.372636] env[61986]: value = "task-1159773" [ 861.372636] env[61986]: _type = "Task" [ 861.372636] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.384294] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159773, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.395040] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159772, 'name': CreateVM_Task, 'duration_secs': 0.879718} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.395040] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 861.401326] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.401444] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.401777] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.403595] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32589566-fc50-4807-b333-29e89a8cc866 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.408548] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 861.408548] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5292d9bd-363c-c807-8278-806ae683cf04" [ 861.408548] env[61986]: _type = "Task" [ 861.408548] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.419584] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5292d9bd-363c-c807-8278-806ae683cf04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.445708] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Updating instance_info_cache with network_info: [{"id": "18e4cc01-51d6-4a7f-b279-4427cedb88ca", "address": "fa:16:3e:72:c3:bf", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18e4cc01-51", "ovs_interfaceid": "18e4cc01-51d6-4a7f-b279-4427cedb88ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.641022] env[61986]: WARNING nova.network.neutron [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] 44bde880-8646-44c4-8c3c-c1faf7d4a9ce already exists in list: networks containing: ['44bde880-8646-44c4-8c3c-c1faf7d4a9ce']. ignoring it [ 861.641022] env[61986]: WARNING nova.network.neutron [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] 44bde880-8646-44c4-8c3c-c1faf7d4a9ce already exists in list: networks containing: ['44bde880-8646-44c4-8c3c-c1faf7d4a9ce']. ignoring it [ 861.655525] env[61986]: DEBUG nova.compute.utils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.657217] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.657486] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 861.667024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ec29ac3e-ffd8-475e-bd38-581214ec8e33 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.847s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.699147] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cc70df1d-d829-42a5-81fd-823a9b178665 tempest-ServerPasswordTestJSON-1650180091 tempest-ServerPasswordTestJSON-1650180091-project-member] Lock "ebf5fee5-7f1b-4537-aec3-77a8a963670b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.733s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.765470] env[61986]: DEBUG nova.policy [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21b078b477724e9581eba5760ddbd6bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4a585cde0f849f8929d526014ed84d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 861.882927] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159773, 'name': Rename_Task, 'duration_secs': 0.136057} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.885527] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 861.885889] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-525c1815-4fe0-47ee-85a9-d6d897a208e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.892616] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 861.892616] env[61986]: value = "task-1159774" [ 861.892616] env[61986]: _type = "Task" [ 861.892616] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.900295] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.918427] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5292d9bd-363c-c807-8278-806ae683cf04, 'name': SearchDatastore_Task, 'duration_secs': 0.010427} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.919117] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.919520] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.919895] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.922018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.922018] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.922018] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4587cbae-9173-40fd-a902-0f1512eda577 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.932110] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.932110] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 861.932110] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35a4aa08-0ef3-432a-a7b9-2356e82bfe85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.937763] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 861.937763] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fdaf7c-852e-0c66-b1a5-aa9cc5496725" [ 861.937763] env[61986]: _type = "Task" [ 861.937763] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.948145] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fdaf7c-852e-0c66-b1a5-aa9cc5496725, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.950843] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "refresh_cache-076c8e91-af70-489f-89d5-f598166fafc6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.951297] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Instance network_info: |[{"id": "18e4cc01-51d6-4a7f-b279-4427cedb88ca", "address": "fa:16:3e:72:c3:bf", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18e4cc01-51", "ovs_interfaceid": "18e4cc01-51d6-4a7f-b279-4427cedb88ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.951885] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:c3:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18e4cc01-51d6-4a7f-b279-4427cedb88ca', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.962042] env[61986]: DEBUG oslo.service.loopingcall [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.962042] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 861.962042] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22d14c7f-3057-4844-b56b-2f5fd468f433 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.983042] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.983042] env[61986]: value = "task-1159775" [ 861.983042] env[61986]: _type = "Task" [ 861.983042] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.988339] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159775, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.162180] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.173020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-588e41c3-aace-4e4a-8f29-c736d5706830 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.173020] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.178714] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 23.701s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.187262] env[61986]: INFO nova.compute.manager [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Rescuing [ 862.187356] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.187485] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.187669] env[61986]: DEBUG nova.network.neutron [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.256032] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Successfully created port: 2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.380484] env[61986]: DEBUG nova.network.neutron [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b2150478-f9cd-4dba-94fe-205081032d36", "address": "fa:16:3e:15:6a:9c", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2150478-f9", "ovs_interfaceid": "b2150478-f9cd-4dba-94fe-205081032d36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5a6caae3-e994-43ff-b2b4-5e3f89aea659", "address": "fa:16:3e:da:fe:6d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a6caae3-e9", "ovs_interfaceid": "5a6caae3-e994-43ff-b2b4-5e3f89aea659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.406681] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159774, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.448570] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fdaf7c-852e-0c66-b1a5-aa9cc5496725, 'name': SearchDatastore_Task, 'duration_secs': 0.008282} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.449677] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-556a89e4-c5b2-41e4-a944-75964bb57a58 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.455385] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 862.455385] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d13b8b-95d8-6717-b1ad-653006abaacd" [ 862.455385] env[61986]: _type = "Task" [ 862.455385] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.463278] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d13b8b-95d8-6717-b1ad-653006abaacd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.492480] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159775, 'name': CreateVM_Task, 'duration_secs': 0.480493} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.492724] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.494213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.494213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.494213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.494456] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77a586bd-05cb-4035-a322-2ea3185b5356 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.499831] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 862.499831] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f0ae78-44b4-105a-43c8-554d370f32ae" [ 862.499831] env[61986]: _type = "Task" [ 862.499831] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.507928] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f0ae78-44b4-105a-43c8-554d370f32ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.682992] env[61986]: DEBUG nova.objects.instance [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lazy-loading 'migration_context' on Instance uuid 40de5f22-0de2-466a-91ab-dcb6ec586dad {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.712629] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.886650] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.887404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.887572] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.889215] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b602b7a-9889-484e-82cf-4e4bc55b1144 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.909411] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.909948] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.910050] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.910241] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.910392] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.910560] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.910741] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.910901] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.911086] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.911259] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.911434] env[61986]: DEBUG nova.virt.hardware [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.919348] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfiguring VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 862.926976] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-993c162c-b744-4bec-bf4d-d08d2d24a365 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.950614] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159774, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.952126] env[61986]: DEBUG oslo_vmware.api [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 862.952126] env[61986]: value = "task-1159776" [ 862.952126] env[61986]: _type = "Task" [ 862.952126] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.967258] env[61986]: DEBUG oslo_vmware.api [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159776, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.973967] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d13b8b-95d8-6717-b1ad-653006abaacd, 'name': SearchDatastore_Task, 'duration_secs': 0.010868} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.974291] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.974574] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 46b18b6c-6e40-45b7-9d3f-6177b08a52a5/46b18b6c-6e40-45b7-9d3f-6177b08a52a5.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 862.974865] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c981c1c7-e075-48da-9422-63767860f6b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.982410] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 862.982410] env[61986]: value = "task-1159777" [ 862.982410] env[61986]: _type = "Task" [ 862.982410] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.994551] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.011166] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f0ae78-44b4-105a-43c8-554d370f32ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009885} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.011391] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.011629] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.011857] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.012036] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.012244] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.012523] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0b703f3-2da4-48cc-ac58-430c007b9d93 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.020989] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.021448] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 863.022908] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a365ae8-6538-4a79-a1b2-2819080b7d9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.027738] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 863.027738] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528677ce-4905-d408-c0c5-bad6dc1c850e" [ 863.027738] env[61986]: _type = "Task" [ 863.027738] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.035822] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528677ce-4905-d408-c0c5-bad6dc1c850e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.063850] env[61986]: DEBUG nova.network.neutron [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Updating instance_info_cache with network_info: [{"id": "9456eeae-0969-4fb6-b614-a699facd11f9", "address": "fa:16:3e:ca:f0:1b", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9456eeae-09", "ovs_interfaceid": "9456eeae-0969-4fb6-b614-a699facd11f9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.175483] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.190901] env[61986]: DEBUG nova.compute.manager [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Received event network-changed-18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 863.191086] env[61986]: DEBUG nova.compute.manager [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Refreshing instance network info cache due to event network-changed-18e4cc01-51d6-4a7f-b279-4427cedb88ca. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 863.191314] env[61986]: DEBUG oslo_concurrency.lockutils [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] Acquiring lock "refresh_cache-076c8e91-af70-489f-89d5-f598166fafc6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.191455] env[61986]: DEBUG oslo_concurrency.lockutils [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] Acquired lock "refresh_cache-076c8e91-af70-489f-89d5-f598166fafc6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.191681] env[61986]: DEBUG nova.network.neutron [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Refreshing network info cache for port 18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.208411] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.208670] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.208831] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.209059] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.209212] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.209363] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.209571] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.209732] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.209905] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.210274] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.210505] env[61986]: DEBUG nova.virt.hardware [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.211628] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286cb0c7-bb18-467a-9f9d-a62eaa5040f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.221028] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016022c4-6b1e-4abe-9180-8aefccaea899 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.310358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.310358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.310358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.310358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.310358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.313539] env[61986]: INFO nova.compute.manager [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Terminating instance [ 863.319822] env[61986]: DEBUG nova.compute.manager [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.320069] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 863.321076] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef31fda-e123-493e-aef3-087c08cd8418 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.329539] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 863.330381] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93df8e10-b93e-4374-8ddb-10b222ec7b0d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.341596] env[61986]: DEBUG oslo_vmware.api [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 863.341596] env[61986]: value = "task-1159778" [ 863.341596] env[61986]: _type = "Task" [ 863.341596] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.359801] env[61986]: DEBUG oslo_vmware.api [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.419228] env[61986]: DEBUG oslo_vmware.api [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159774, 'name': PowerOnVM_Task, 'duration_secs': 1.325177} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.422585] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 863.422964] env[61986]: INFO nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Took 9.68 seconds to spawn the instance on the hypervisor. [ 863.423264] env[61986]: DEBUG nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.424615] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63da303e-f6b6-4e55-b2fd-c2528c8f9525 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.466066] env[61986]: DEBUG oslo_vmware.api [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159776, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.500023] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159777, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.541828] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528677ce-4905-d408-c0c5-bad6dc1c850e, 'name': SearchDatastore_Task, 'duration_secs': 0.011936} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.546558] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd9c6c74-9350-44c7-9022-d034a62383b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.555564] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 863.555564] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed46ae-8647-242b-4c30-a1ec72932725" [ 863.555564] env[61986]: _type = "Task" [ 863.555564] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.568654] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "refresh_cache-021a27ec-cee0-454d-8daf-e6a82bd9330c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.571020] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed46ae-8647-242b-4c30-a1ec72932725, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.785989] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18031ec6-e023-4805-8ec9-adf5b043557f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.793698] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340e409e-18ed-4488-87c0-dc8df144889f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.828430] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8158629c-2839-4e1a-89e0-b15dc7f38bf2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.836386] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7ed2b0-4b69-461d-a346-a9e0097bcf66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.854551] env[61986]: DEBUG nova.compute.provider_tree [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.861119] env[61986]: DEBUG oslo_vmware.api [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159778, 'name': PowerOffVM_Task, 'duration_secs': 0.282589} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.861362] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.861600] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.862416] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df70299d-73da-4266-8635-27a58ca6d432 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.959567] env[61986]: INFO nova.compute.manager [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Took 44.59 seconds to build instance. [ 863.966606] env[61986]: DEBUG oslo_vmware.api [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159776, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.975470] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.975696] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.975870] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Deleting the datastore file [datastore1] 6f38006d-b1f5-4aeb-9124-1e32378c22b2 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.976240] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43fce7e4-3664-4f05-b965-ab09d3327dd2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.984745] env[61986]: DEBUG oslo_vmware.api [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for the task: (returnval){ [ 863.984745] env[61986]: value = "task-1159780" [ 863.984745] env[61986]: _type = "Task" [ 863.984745] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.000800] env[61986]: DEBUG oslo_vmware.api [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.002577] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159777, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.731785} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.002861] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 46b18b6c-6e40-45b7-9d3f-6177b08a52a5/46b18b6c-6e40-45b7-9d3f-6177b08a52a5.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 864.003288] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.003553] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c1eb0fd-e998-426c-b045-5ce339686f1b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.010524] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 864.010524] env[61986]: value = "task-1159781" [ 864.010524] env[61986]: _type = "Task" [ 864.010524] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.019989] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159781, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.069619] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed46ae-8647-242b-4c30-a1ec72932725, 'name': SearchDatastore_Task, 'duration_secs': 0.060162} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.069922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.070792] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 076c8e91-af70-489f-89d5-f598166fafc6/076c8e91-af70-489f-89d5-f598166fafc6.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 864.070792] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11776aa9-65f2-4cdb-a96c-55bd0690c8e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.078757] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 864.078757] env[61986]: value = "task-1159782" [ 864.078757] env[61986]: _type = "Task" [ 864.078757] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.088085] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159782, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.121530] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.121813] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a95bf7d6-b3fd-44b4-82ab-e8b795a8dbdc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.129616] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 864.129616] env[61986]: value = "task-1159783" [ 864.129616] env[61986]: _type = "Task" [ 864.129616] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.137846] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.189121] env[61986]: DEBUG nova.network.neutron [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Updated VIF entry in instance network info cache for port 18e4cc01-51d6-4a7f-b279-4427cedb88ca. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.189121] env[61986]: DEBUG nova.network.neutron [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Updating instance_info_cache with network_info: [{"id": "18e4cc01-51d6-4a7f-b279-4427cedb88ca", "address": "fa:16:3e:72:c3:bf", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18e4cc01-51", "ovs_interfaceid": "18e4cc01-51d6-4a7f-b279-4427cedb88ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.317754] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Successfully updated port: 2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.358564] env[61986]: DEBUG nova.scheduler.client.report [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.462288] env[61986]: DEBUG oslo_concurrency.lockutils [None req-47c88ea8-935b-4ab5-99a2-e2b05c380f99 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.044s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.473418] env[61986]: DEBUG oslo_vmware.api [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159776, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.499129] env[61986]: DEBUG oslo_vmware.api [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Task: {'id': task-1159780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326004} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.499437] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.499632] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.499813] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.500071] env[61986]: INFO nova.compute.manager [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Took 1.18 seconds to destroy the instance on the hypervisor. [ 864.500345] env[61986]: DEBUG oslo.service.loopingcall [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.500578] env[61986]: DEBUG nova.compute.manager [-] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.500687] env[61986]: DEBUG nova.network.neutron [-] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.523958] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159781, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155247} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.524274] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.525096] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d913a6aa-e4ed-4f90-99b7-eb5ec3fe7ce6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.551250] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 46b18b6c-6e40-45b7-9d3f-6177b08a52a5/46b18b6c-6e40-45b7-9d3f-6177b08a52a5.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.551376] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fe3b565-7622-4b0f-8544-d82971c5e867 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.572559] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 864.572559] env[61986]: value = "task-1159785" [ 864.572559] env[61986]: _type = "Task" [ 864.572559] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.583996] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.592953] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159782, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.641239] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159783, 'name': PowerOffVM_Task, 'duration_secs': 0.211837} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.641578] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 864.642450] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f01e3b4-fb3d-4cc8-8de4-b0b8dabc23c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.662591] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1550186-1d72-4a59-a6cd-e838aa1723b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.691991] env[61986]: DEBUG oslo_concurrency.lockutils [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] Releasing lock "refresh_cache-076c8e91-af70-489f-89d5-f598166fafc6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.692275] env[61986]: DEBUG nova.compute.manager [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-changed-5a6caae3-e994-43ff-b2b4-5e3f89aea659 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.692451] env[61986]: DEBUG nova.compute.manager [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing instance network info cache due to event network-changed-5a6caae3-e994-43ff-b2b4-5e3f89aea659. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.692661] env[61986]: DEBUG oslo_concurrency.lockutils [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.692806] env[61986]: DEBUG oslo_concurrency.lockutils [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.692966] env[61986]: DEBUG nova.network.neutron [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Refreshing network info cache for port 5a6caae3-e994-43ff-b2b4-5e3f89aea659 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.703225] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.703506] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f53afdfb-b90f-4e52-afc7-6463f87dee25 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.711039] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 864.711039] env[61986]: value = "task-1159788" [ 864.711039] env[61986]: _type = "Task" [ 864.711039] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.718885] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159788, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.824606] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "refresh_cache-52eda388-05f4-416c-ac9d-bd1c1e31ba9c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.824645] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "refresh_cache-52eda388-05f4-416c-ac9d-bd1c1e31ba9c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.824828] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.971351] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.973887] env[61986]: DEBUG oslo_vmware.api [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159776, 'name': ReconfigVM_Task, 'duration_secs': 1.555643} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.974823] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.974823] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfigured VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 865.085865] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159785, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.098503] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159782, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600967} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.099063] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 076c8e91-af70-489f-89d5-f598166fafc6/076c8e91-af70-489f-89d5-f598166fafc6.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.099182] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.099391] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65c7db37-67d8-4f98-b732-b38b8905a0a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.107090] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 865.107090] env[61986]: value = "task-1159789" [ 865.107090] env[61986]: _type = "Task" [ 865.107090] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.118313] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.227500] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 865.227847] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.227998] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.228190] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.228377] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.228715] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f26439bd-baaf-41f1-8a03-667c05bfd627 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.241894] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.242109] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 865.242865] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c195bb92-f6c1-4152-9be8-807b4a006450 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.248554] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 865.248554] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52769ef1-9b03-f02d-a7bf-1af0e5d130f5" [ 865.248554] env[61986]: _type = "Task" [ 865.248554] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.261961] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52769ef1-9b03-f02d-a7bf-1af0e5d130f5, 'name': SearchDatastore_Task, 'duration_secs': 0.008555} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.262702] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9a098e0-ed31-4fec-9410-c8a8617f80f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.268415] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 865.268415] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52334724-96f7-05f7-ebf6-0f92a356f0f4" [ 865.268415] env[61986]: _type = "Task" [ 865.268415] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.279641] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52334724-96f7-05f7-ebf6-0f92a356f0f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.288157] env[61986]: DEBUG nova.network.neutron [-] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.375957] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.199s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.385291] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.387450] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.264s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.389047] env[61986]: INFO nova.compute.claims [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.479707] env[61986]: DEBUG oslo_concurrency.lockutils [None req-27810aa6-2c73-44c4-9738-9865d5638c16 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-5a6caae3-e994-43ff-b2b4-5e3f89aea659" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.702s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.502209] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.590533] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159785, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.594147] env[61986]: DEBUG nova.network.neutron [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updated VIF entry in instance network info cache for port 5a6caae3-e994-43ff-b2b4-5e3f89aea659. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.595274] env[61986]: DEBUG nova.network.neutron [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b2150478-f9cd-4dba-94fe-205081032d36", "address": "fa:16:3e:15:6a:9c", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2150478-f9", "ovs_interfaceid": "b2150478-f9cd-4dba-94fe-205081032d36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5a6caae3-e994-43ff-b2b4-5e3f89aea659", "address": "fa:16:3e:da:fe:6d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a6caae3-e9", "ovs_interfaceid": "5a6caae3-e994-43ff-b2b4-5e3f89aea659", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.618957] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067729} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.619063] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.620123] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2bc270-e53b-49ca-b47d-afb7d983a2d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.644194] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 076c8e91-af70-489f-89d5-f598166fafc6/076c8e91-af70-489f-89d5-f598166fafc6.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.645368] env[61986]: DEBUG nova.network.neutron [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Updating instance_info_cache with network_info: [{"id": "2deaa3c0-1890-4e85-83f6-1ca9a7007616", "address": "fa:16:3e:34:58:12", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deaa3c0-18", "ovs_interfaceid": "2deaa3c0-1890-4e85-83f6-1ca9a7007616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.646941] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-309fed60-c952-4c51-abad-84a07c0ce979 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.667738] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 865.667738] env[61986]: value = "task-1159790" [ 865.667738] env[61986]: _type = "Task" [ 865.667738] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.676108] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159790, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.781692] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52334724-96f7-05f7-ebf6-0f92a356f0f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009293} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.782020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.782573] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk. {{(pid=61986) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 865.782703] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a823ec0-5049-4a7e-91fd-1e2b491726c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.790480] env[61986]: INFO nova.compute.manager [-] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Took 1.29 seconds to deallocate network for instance. [ 865.791887] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 865.791887] env[61986]: value = "task-1159791" [ 865.791887] env[61986]: _type = "Task" [ 865.791887] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.805979] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.841052] env[61986]: DEBUG nova.compute.manager [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Received event network-vif-plugged-2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.841305] env[61986]: DEBUG oslo_concurrency.lockutils [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] Acquiring lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.841547] env[61986]: DEBUG oslo_concurrency.lockutils [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.841767] env[61986]: DEBUG oslo_concurrency.lockutils [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.841970] env[61986]: DEBUG nova.compute.manager [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] No waiting events found dispatching network-vif-plugged-2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.842219] env[61986]: WARNING nova.compute.manager [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Received unexpected event network-vif-plugged-2deaa3c0-1890-4e85-83f6-1ca9a7007616 for instance with vm_state building and task_state spawning. [ 865.842433] env[61986]: DEBUG nova.compute.manager [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Received event network-changed-2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.842611] env[61986]: DEBUG nova.compute.manager [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Refreshing instance network info cache due to event network-changed-2deaa3c0-1890-4e85-83f6-1ca9a7007616. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 865.842816] env[61986]: DEBUG oslo_concurrency.lockutils [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] Acquiring lock "refresh_cache-52eda388-05f4-416c-ac9d-bd1c1e31ba9c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.085747] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159785, 'name': ReconfigVM_Task, 'duration_secs': 1.469698} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.086147] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 46b18b6c-6e40-45b7-9d3f-6177b08a52a5/46b18b6c-6e40-45b7-9d3f-6177b08a52a5.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.087040] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa917a33-bc93-417f-b02f-31bb6e72487a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.094623] env[61986]: DEBUG nova.compute.manager [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Received event network-changed-3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.094793] env[61986]: DEBUG nova.compute.manager [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Refreshing instance network info cache due to event network-changed-3cf0a017-8fc3-4649-bb5a-922ad76c9140. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 866.094996] env[61986]: DEBUG oslo_concurrency.lockutils [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] Acquiring lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.095234] env[61986]: DEBUG oslo_concurrency.lockutils [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] Acquired lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.095348] env[61986]: DEBUG nova.network.neutron [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Refreshing network info cache for port 3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.098874] env[61986]: DEBUG oslo_concurrency.lockutils [req-446e9d66-678d-4b6d-9ea7-f0dd4fc0e427 req-79e30ae7-d0a2-4035-9034-cbe572bec5bc service nova] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.099353] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 866.099353] env[61986]: value = "task-1159792" [ 866.099353] env[61986]: _type = "Task" [ 866.099353] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.110023] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159792, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.162305] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "refresh_cache-52eda388-05f4-416c-ac9d-bd1c1e31ba9c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.162729] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Instance network_info: |[{"id": "2deaa3c0-1890-4e85-83f6-1ca9a7007616", "address": "fa:16:3e:34:58:12", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deaa3c0-18", "ovs_interfaceid": "2deaa3c0-1890-4e85-83f6-1ca9a7007616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.163100] env[61986]: DEBUG oslo_concurrency.lockutils [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] Acquired lock "refresh_cache-52eda388-05f4-416c-ac9d-bd1c1e31ba9c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.163326] env[61986]: DEBUG nova.network.neutron [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Refreshing network info cache for port 2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.164757] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:58:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2deaa3c0-1890-4e85-83f6-1ca9a7007616', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.173311] env[61986]: DEBUG oslo.service.loopingcall [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.174339] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 866.179263] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c247953-74a9-4b9d-9105-3b6409b31da6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.200307] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159790, 'name': ReconfigVM_Task, 'duration_secs': 0.305668} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.200645] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 076c8e91-af70-489f-89d5-f598166fafc6/076c8e91-af70-489f-89d5-f598166fafc6.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.202928] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6aa10de-96d0-4672-8002-46c41798752c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.204650] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.204650] env[61986]: value = "task-1159793" [ 866.204650] env[61986]: _type = "Task" [ 866.204650] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.210923] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 866.210923] env[61986]: value = "task-1159794" [ 866.210923] env[61986]: _type = "Task" [ 866.210923] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.217856] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159793, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.223715] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159794, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.301718] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.306512] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159791, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.613245] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159792, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.716417] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159793, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.732500] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159794, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.803915] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159791, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561403} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.804232] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk. [ 866.805036] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4760391d-17c1-421e-9e78-c2dc659a0dfc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.835478] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.842658] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0aa6add-146b-4429-ae1a-5a8ebdd5cf41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.867907] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 866.867907] env[61986]: value = "task-1159795" [ 866.867907] env[61986]: _type = "Task" [ 866.867907] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.886373] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159795, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.939959] env[61986]: INFO nova.compute.manager [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Swapping old allocation on dict_keys(['2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d']) held by migration 84ab935e-920f-44a1-8d8e-f5b0438906b8 for instance [ 866.981534] env[61986]: DEBUG nova.scheduler.client.report [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Overwriting current allocation {'allocations': {'2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 85}}, 'project_id': 'ddd77efba5d94caf81e14827bbae6ed8', 'user_id': 'f3ca3baaab63400684218e32dc8b8bdf', 'consumer_generation': 1} on consumer 40de5f22-0de2-466a-91ab-dcb6ec586dad {{(pid=61986) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 867.007038] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccff2fdb-cce6-4f58-86f5-5761aa468762 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.012543] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2460f8-0abc-4235-9a6b-9682b3bf204f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.044162] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0fbb44-588d-4b7c-aba0-b86c35b17667 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.057266] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cf5a1d-6beb-4dc8-a2c5-cbf98a8916d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.066212] env[61986]: DEBUG nova.network.neutron [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updated VIF entry in instance network info cache for port 3cf0a017-8fc3-4649-bb5a-922ad76c9140. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.066212] env[61986]: DEBUG nova.network.neutron [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating instance_info_cache with network_info: [{"id": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "address": "fa:16:3e:80:d7:ec", "network": {"id": "dde733ac-b7c6-4939-a972-c681f487c6b8", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1722442317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30cffc57b42847d8a13ac2347ea2ade7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cf0a017-8f", "ovs_interfaceid": "3cf0a017-8fc3-4649-bb5a-922ad76c9140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.076966] env[61986]: DEBUG nova.compute.provider_tree [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.097330] env[61986]: DEBUG nova.network.neutron [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Updated VIF entry in instance network info cache for port 2deaa3c0-1890-4e85-83f6-1ca9a7007616. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.097633] env[61986]: DEBUG nova.network.neutron [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Updating instance_info_cache with network_info: [{"id": "2deaa3c0-1890-4e85-83f6-1ca9a7007616", "address": "fa:16:3e:34:58:12", "network": {"id": "e61035ac-52a9-4ce5-bfee-af5bce22290d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1413150173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4a585cde0f849f8929d526014ed84d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deaa3c0-18", "ovs_interfaceid": "2deaa3c0-1890-4e85-83f6-1ca9a7007616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.111189] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159792, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.152397] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.152598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquired lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.152783] env[61986]: DEBUG nova.network.neutron [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 867.216768] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159793, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.225158] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159794, 'name': Rename_Task, 'duration_secs': 0.842733} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.225456] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.225709] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbdc44af-9486-4409-af3a-9b446bce194f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.234041] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 867.234041] env[61986]: value = "task-1159796" [ 867.234041] env[61986]: _type = "Task" [ 867.234041] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.242994] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.379942] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159795, 'name': ReconfigVM_Task, 'duration_secs': 0.320252} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.380306] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.381265] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5d39e9-656e-44dc-ab2b-63532df980eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.409871] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94827e3b-8415-457a-abc2-cc0f4c081696 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.420985] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-977799b4-2793-4513-9447-483146fc7ac4-b2150478-f9cd-4dba-94fe-205081032d36" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.421288] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-b2150478-f9cd-4dba-94fe-205081032d36" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.426581] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 867.426581] env[61986]: value = "task-1159797" [ 867.426581] env[61986]: _type = "Task" [ 867.426581] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.436885] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159797, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.579543] env[61986]: DEBUG oslo_concurrency.lockutils [req-4a699bdd-81ae-496b-a31e-7ab217d094e0 req-d72f5daf-33e4-42a1-b072-8e8b1cac14eb service nova] Releasing lock "refresh_cache-a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.580714] env[61986]: DEBUG nova.scheduler.client.report [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.600113] env[61986]: DEBUG oslo_concurrency.lockutils [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] Releasing lock "refresh_cache-52eda388-05f4-416c-ac9d-bd1c1e31ba9c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.600518] env[61986]: DEBUG nova.compute.manager [req-5afb8a5d-70fc-4385-ba34-c68570b675c6 req-b58a397a-30a1-4244-8b1f-7e11d1a3c828 service nova] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Received event network-vif-deleted-1c018a8d-09c1-4372-b4fd-f35561ab799b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.612977] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159792, 'name': Rename_Task, 'duration_secs': 1.209838} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.613293] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.613546] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b32332f-9e3d-44b1-84f7-513d074cd959 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.619179] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 867.619179] env[61986]: value = "task-1159799" [ 867.619179] env[61986]: _type = "Task" [ 867.619179] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.626714] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159799, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.720476] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159793, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.744587] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159796, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.924099] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.924465] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.925795] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bc8d24-aade-450a-8c37-c8e6525136a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.941076] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159797, 'name': ReconfigVM_Task, 'duration_secs': 0.34584} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.952442] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.953172] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3cedd6a1-6831-41ac-ab66-f6d12bc8836a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.956846] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010052a1-1c3a-44b6-9488-be7cb9e18add {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.989105] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfiguring VM to detach interface {{(pid=61986) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 867.992051] env[61986]: DEBUG nova.network.neutron [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [{"id": "dd125b45-dffb-4a25-af78-4a9326408be0", "address": "fa:16:3e:df:7e:b4", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.98", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd125b45-df", "ovs_interfaceid": "dd125b45-dffb-4a25-af78-4a9326408be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.993448] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9c90b5c-a91f-4c73-ac6e-286225f3b0bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.007380] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 868.007380] env[61986]: value = "task-1159800" [ 868.007380] env[61986]: _type = "Task" [ 868.007380] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.015275] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 868.015275] env[61986]: value = "task-1159801" [ 868.015275] env[61986]: _type = "Task" [ 868.015275] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.025257] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159800, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.031623] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.059813] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "dea128a9-aed0-40b2-ae17-c068ea8e3452" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.060745] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "dea128a9-aed0-40b2-ae17-c068ea8e3452" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.085520] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.698s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.086012] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.088861] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.710s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.090385] env[61986]: INFO nova.compute.claims [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.129785] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159799, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.218622] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159793, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.244859] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159796, 'name': PowerOnVM_Task, 'duration_secs': 0.81894} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.245801] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 868.245801] env[61986]: INFO nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Took 9.01 seconds to spawn the instance on the hypervisor. [ 868.245801] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.247040] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355dab3e-e454-4997-8f8f-5fe0c4844b39 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.509582] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Releasing lock "refresh_cache-40de5f22-0de2-466a-91ab-dcb6ec586dad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.510206] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.510549] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ba167f9-7cde-455d-a3fc-0985315d66e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.521723] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 868.521723] env[61986]: value = "task-1159802" [ 868.521723] env[61986]: _type = "Task" [ 868.521723] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.528661] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.528971] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159800, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.536553] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.595300] env[61986]: DEBUG nova.compute.utils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.599099] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.599281] env[61986]: DEBUG nova.network.neutron [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 868.631728] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159799, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.672371] env[61986]: DEBUG nova.policy [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4675c2084b6463bb47c93865f5cec46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f746f57f466f480c958d8b15bd2ce8fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 868.716656] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159793, 'name': CreateVM_Task, 'duration_secs': 2.494711} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.716818] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 868.717514] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.717682] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.718031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.718316] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0a0c0bb-d3fb-4be0-8c68-2e4237891b20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.722722] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 868.722722] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cf3b47-2c66-6b85-ee83-4a691c4dd92d" [ 868.722722] env[61986]: _type = "Task" [ 868.722722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.730205] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cf3b47-2c66-6b85-ee83-4a691c4dd92d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.763867] env[61986]: INFO nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Took 40.14 seconds to build instance. [ 869.023121] env[61986]: DEBUG oslo_vmware.api [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159800, 'name': PowerOnVM_Task, 'duration_secs': 0.735095} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.023803] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.028301] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.029234] env[61986]: DEBUG nova.compute.manager [None req-5f0bcf9c-7241-4474-b70d-629a28d9ffc7 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.030175] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3177a1d3-8934-4654-adfa-a5ac883cd082 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.044313] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159802, 'name': PowerOffVM_Task, 'duration_secs': 0.457642} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.044313] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:40:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='754ce345-2135-4db7-a907-0c2bf294269e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2075393374',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.046724] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.052132] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23edde5f-e613-4049-81d2-1cadf5bb93c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.074685] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 869.074685] env[61986]: value = "task-1159803" [ 869.074685] env[61986]: _type = "Task" [ 869.074685] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.084361] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159803, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.092929] env[61986]: DEBUG nova.network.neutron [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Successfully created port: 1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.100155] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.135649] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159799, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.232134] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cf3b47-2c66-6b85-ee83-4a691c4dd92d, 'name': SearchDatastore_Task, 'duration_secs': 0.036515} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.234729] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.234971] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.235220] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.235368] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.235544] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.235975] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-335cdd65-f7c2-4211-9fea-27287d61a231 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.248812] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.248812] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 869.249200] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ee37515-514c-45c7-9c44-f30b96e11de2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.254188] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 869.254188] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c88927-9b5c-e044-8786-e832fe3243c1" [ 869.254188] env[61986]: _type = "Task" [ 869.254188] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.265904] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "076c8e91-af70-489f-89d5-f598166fafc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.302s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.266357] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c88927-9b5c-e044-8786-e832fe3243c1, 'name': SearchDatastore_Task, 'duration_secs': 0.008995} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.268125] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccec1292-89ea-4e29-a2f1-f8804536a555 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.278763] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 869.278763] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521b4c04-50dd-4dd0-637b-0eec47ba6ff4" [ 869.278763] env[61986]: _type = "Task" [ 869.278763] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.286663] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521b4c04-50dd-4dd0-637b-0eec47ba6ff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.528442] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.559189] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f925f15f-5ce4-4626-a90c-d9fcceac740f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.568207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36543db0-a323-4215-8d79-234aedfaf7f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.605037] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca03c79-2379-4fc8-8866-22bc2b1ba72e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.616790] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159803, 'name': ReconfigVM_Task, 'duration_secs': 0.289633} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.620028] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c391264-730e-44a3-91f0-a082809ca97b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.623547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e667bc-19c9-43e7-8916-37cfafa79986 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.641831] env[61986]: DEBUG nova.compute.provider_tree [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.658596] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159799, 'name': PowerOnVM_Task, 'duration_secs': 1.952852} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.659554] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:40:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='754ce345-2135-4db7-a907-0c2bf294269e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2075393374',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.659814] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.660016] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.660276] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.660547] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.660820] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.661123] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.661343] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.661603] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.661767] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.661951] env[61986]: DEBUG nova.virt.hardware [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.664526] env[61986]: DEBUG nova.scheduler.client.report [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.667726] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.667831] env[61986]: INFO nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Took 13.21 seconds to spawn the instance on the hypervisor. [ 869.668036] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.668327] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e89c56a-e644-4cce-b2a4-203ae477d884 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.673305] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a947160-3d68-4f01-89a2-30b62ed6f801 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.679432] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 869.679432] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e2647-b5ba-77dd-43f3-cbfacf4dc65b" [ 869.679432] env[61986]: _type = "Task" [ 869.679432] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.697543] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e2647-b5ba-77dd-43f3-cbfacf4dc65b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.772017] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.788270] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521b4c04-50dd-4dd0-637b-0eec47ba6ff4, 'name': SearchDatastore_Task, 'duration_secs': 0.010918} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.788546] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.788806] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 52eda388-05f4-416c-ac9d-bd1c1e31ba9c/52eda388-05f4-416c-ac9d-bd1c1e31ba9c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 869.789113] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a5a4420-3d3b-4823-812a-fd4aeacf5aa2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.796327] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 869.796327] env[61986]: value = "task-1159805" [ 869.796327] env[61986]: _type = "Task" [ 869.796327] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.803787] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159805, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.030360] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.116038] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.142299] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.142588] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.142757] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.142978] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.143110] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.143270] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.143516] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.145149] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.145149] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.145149] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.145149] env[61986]: DEBUG nova.virt.hardware [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.145430] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2c33ef-abbd-4924-a735-4f49293202c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.158757] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e88dc3-1fc9-436d-8e3a-23f3ca41171d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.174114] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.085s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.174324] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.177091] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.686s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.178683] env[61986]: INFO nova.compute.claims [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.195958] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e2647-b5ba-77dd-43f3-cbfacf4dc65b, 'name': SearchDatastore_Task, 'duration_secs': 0.016437} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.203395] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfiguring VM instance instance-00000029 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 870.204067] env[61986]: INFO nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Took 44.31 seconds to build instance. [ 870.205090] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83517cd3-08dd-4811-9726-9fd9baeba504 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.229800] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 870.229800] env[61986]: value = "task-1159806" [ 870.229800] env[61986]: _type = "Task" [ 870.229800] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.238698] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159806, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.294105] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.308375] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159805, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.505812] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.506031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.527148] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.613820] env[61986]: DEBUG nova.compute.manager [req-ad3c7a30-63e3-45be-b78b-81c4f89a16e1 req-254541f4-ea36-41b6-aa23-7b07940fb236 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Received event network-vif-plugged-1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.614129] env[61986]: DEBUG oslo_concurrency.lockutils [req-ad3c7a30-63e3-45be-b78b-81c4f89a16e1 req-254541f4-ea36-41b6-aa23-7b07940fb236 service nova] Acquiring lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.614716] env[61986]: DEBUG oslo_concurrency.lockutils [req-ad3c7a30-63e3-45be-b78b-81c4f89a16e1 req-254541f4-ea36-41b6-aa23-7b07940fb236 service nova] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.614877] env[61986]: DEBUG oslo_concurrency.lockutils [req-ad3c7a30-63e3-45be-b78b-81c4f89a16e1 req-254541f4-ea36-41b6-aa23-7b07940fb236 service nova] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.615086] env[61986]: DEBUG nova.compute.manager [req-ad3c7a30-63e3-45be-b78b-81c4f89a16e1 req-254541f4-ea36-41b6-aa23-7b07940fb236 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] No waiting events found dispatching network-vif-plugged-1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.615286] env[61986]: WARNING nova.compute.manager [req-ad3c7a30-63e3-45be-b78b-81c4f89a16e1 req-254541f4-ea36-41b6-aa23-7b07940fb236 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Received unexpected event network-vif-plugged-1234061c-4805-4e9a-9993-3e09549f5569 for instance with vm_state building and task_state spawning. [ 870.683390] env[61986]: DEBUG nova.compute.utils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.686891] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.687099] env[61986]: DEBUG nova.network.neutron [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 870.721088] env[61986]: DEBUG nova.network.neutron [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Successfully updated port: 1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.723186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.787s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.740358] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159806, 'name': ReconfigVM_Task, 'duration_secs': 0.226474} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.740971] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfigured VM instance instance-00000029 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 870.741502] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1377cd01-384b-45d5-9cd2-7abf0c58a770 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.746629] env[61986]: DEBUG nova.policy [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94bb4fe585bb42339a01194fa2a68bac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '034de7dfeeb145c0a1e90b111c15e36f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 870.765836] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.766140] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25633b71-8bdd-49d4-b84c-ae05f03abf78 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.783791] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 870.783791] env[61986]: value = "task-1159807" [ 870.783791] env[61986]: _type = "Task" [ 870.783791] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.794266] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159807, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.806548] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159805, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610787} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.806848] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 52eda388-05f4-416c-ac9d-bd1c1e31ba9c/52eda388-05f4-416c-ac9d-bd1c1e31ba9c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 870.807090] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.807367] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e9aca24-3d8a-4d18-a307-93febc62eb2b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.814077] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 870.814077] env[61986]: value = "task-1159808" [ 870.814077] env[61986]: _type = "Task" [ 870.814077] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.822013] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.029279] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.034538] env[61986]: DEBUG nova.network.neutron [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Successfully created port: 7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.187817] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 871.224382] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "refresh_cache-b9706e51-0f74-4dbc-aab1-ea640ac78c62" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.224534] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "refresh_cache-b9706e51-0f74-4dbc-aab1-ea640ac78c62" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.224697] env[61986]: DEBUG nova.network.neutron [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.229030] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.299830] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159807, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.324741] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063883} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.325871] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.326811] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300f05d9-5dc5-437b-9556-6175d3025ab9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.351940] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 52eda388-05f4-416c-ac9d-bd1c1e31ba9c/52eda388-05f4-416c-ac9d-bd1c1e31ba9c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.352685] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1473090d-d7c4-434f-93af-bd48455d33f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.375559] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 871.375559] env[61986]: value = "task-1159809" [ 871.375559] env[61986]: _type = "Task" [ 871.375559] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.384985] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.530096] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.721821] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84443525-bb55-4b90-a1a9-fbdecac75525 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.729649] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9ee631-2f90-49ef-b095-8453598873f8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.768321] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.769129] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2da9840-bdae-4fdc-86e8-798cf3fb9c9c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.774266] env[61986]: DEBUG nova.network.neutron [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.781026] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bff066f-de25-488f-8fa2-326e23500db6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.795567] env[61986]: DEBUG nova.compute.provider_tree [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.806149] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159807, 'name': ReconfigVM_Task, 'duration_secs': 0.543853} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.806149] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad/40de5f22-0de2-466a-91ab-dcb6ec586dad.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.806149] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7569af13-1024-4e4d-a4be-a895229299b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.823325] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd1d688-3a3f-4bf7-b958-08a91c859057 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.843208] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81df0ba1-abf9-4646-881b-fc8e2480646b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.864918] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210235bb-7417-41e0-aa81-dec196a0234f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.871996] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 871.871996] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a18615b2-f457-4072-a0a1-24ab7c6db3e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.881382] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 871.881382] env[61986]: value = "task-1159811" [ 871.881382] env[61986]: _type = "Task" [ 871.881382] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.886948] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.891400] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.936227] env[61986]: DEBUG nova.network.neutron [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Updating instance_info_cache with network_info: [{"id": "1234061c-4805-4e9a-9993-3e09549f5569", "address": "fa:16:3e:3c:45:e9", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1234061c-48", "ovs_interfaceid": "1234061c-4805-4e9a-9993-3e09549f5569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.032665] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.200122] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.227801] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.228514] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.228788] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.229093] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.229330] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.229629] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.229867] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.230149] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.230375] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.230629] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.230892] env[61986]: DEBUG nova.virt.hardware [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.232139] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26907b86-dddd-49d0-b7ff-908c2c97aee5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.242923] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b0806c-270a-49b4-b9e7-7eb53832e7fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.300586] env[61986]: DEBUG nova.scheduler.client.report [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.389890] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159809, 'name': ReconfigVM_Task, 'duration_secs': 0.793453} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.392874] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 52eda388-05f4-416c-ac9d-bd1c1e31ba9c/52eda388-05f4-416c-ac9d-bd1c1e31ba9c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.393566] env[61986]: DEBUG oslo_vmware.api [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159811, 'name': PowerOnVM_Task, 'duration_secs': 0.420711} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.393776] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd99334e-3cbf-41e3-920d-a8fce470d37b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.395376] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 872.405229] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 872.405229] env[61986]: value = "task-1159812" [ 872.405229] env[61986]: _type = "Task" [ 872.405229] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.413861] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159812, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.438667] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "refresh_cache-b9706e51-0f74-4dbc-aab1-ea640ac78c62" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.439069] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Instance network_info: |[{"id": "1234061c-4805-4e9a-9993-3e09549f5569", "address": "fa:16:3e:3c:45:e9", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1234061c-48", "ovs_interfaceid": "1234061c-4805-4e9a-9993-3e09549f5569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.439546] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:45:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1234061c-4805-4e9a-9993-3e09549f5569', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.447544] env[61986]: DEBUG oslo.service.loopingcall [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.447797] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 872.448049] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f91d532-9d6d-4dd8-9a1f-4f0088e4e11d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.468280] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.468280] env[61986]: value = "task-1159813" [ 872.468280] env[61986]: _type = "Task" [ 872.468280] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.475982] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159813, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.489639] env[61986]: DEBUG nova.compute.manager [req-69bc2e41-5479-4673-a582-03e4c4d593cc req-01b6c49a-cdf7-49fb-85a3-d1fea4a0196e service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Received event network-vif-plugged-7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.489852] env[61986]: DEBUG oslo_concurrency.lockutils [req-69bc2e41-5479-4673-a582-03e4c4d593cc req-01b6c49a-cdf7-49fb-85a3-d1fea4a0196e service nova] Acquiring lock "4f13b629-e2a7-4668-9d77-eb638078e246-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.490087] env[61986]: DEBUG oslo_concurrency.lockutils [req-69bc2e41-5479-4673-a582-03e4c4d593cc req-01b6c49a-cdf7-49fb-85a3-d1fea4a0196e service nova] Lock "4f13b629-e2a7-4668-9d77-eb638078e246-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.490265] env[61986]: DEBUG oslo_concurrency.lockutils [req-69bc2e41-5479-4673-a582-03e4c4d593cc req-01b6c49a-cdf7-49fb-85a3-d1fea4a0196e service nova] Lock "4f13b629-e2a7-4668-9d77-eb638078e246-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.490593] env[61986]: DEBUG nova.compute.manager [req-69bc2e41-5479-4673-a582-03e4c4d593cc req-01b6c49a-cdf7-49fb-85a3-d1fea4a0196e service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] No waiting events found dispatching network-vif-plugged-7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 872.490843] env[61986]: WARNING nova.compute.manager [req-69bc2e41-5479-4673-a582-03e4c4d593cc req-01b6c49a-cdf7-49fb-85a3-d1fea4a0196e service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Received unexpected event network-vif-plugged-7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 for instance with vm_state building and task_state spawning. [ 872.532111] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.577795] env[61986]: DEBUG nova.network.neutron [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Successfully updated port: 7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.639523] env[61986]: DEBUG nova.compute.manager [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Received event network-changed-1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.639733] env[61986]: DEBUG nova.compute.manager [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Refreshing instance network info cache due to event network-changed-1234061c-4805-4e9a-9993-3e09549f5569. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 872.639945] env[61986]: DEBUG oslo_concurrency.lockutils [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] Acquiring lock "refresh_cache-b9706e51-0f74-4dbc-aab1-ea640ac78c62" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.640106] env[61986]: DEBUG oslo_concurrency.lockutils [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] Acquired lock "refresh_cache-b9706e51-0f74-4dbc-aab1-ea640ac78c62" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.640981] env[61986]: DEBUG nova.network.neutron [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Refreshing network info cache for port 1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.807621] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.807621] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.810431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.938s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.810659] env[61986]: DEBUG nova.objects.instance [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lazy-loading 'resources' on Instance uuid 6f20dd8e-9289-458d-bed6-f2ef9daaa917 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.914131] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159812, 'name': Rename_Task, 'duration_secs': 0.205002} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.914394] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.914634] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f03de5e-974a-4b27-ac4c-42311ba52a50 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.921291] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 872.921291] env[61986]: value = "task-1159814" [ 872.921291] env[61986]: _type = "Task" [ 872.921291] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.929076] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159814, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.977888] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159813, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.029672] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.081070] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "refresh_cache-4f13b629-e2a7-4668-9d77-eb638078e246" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.081245] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired lock "refresh_cache-4f13b629-e2a7-4668-9d77-eb638078e246" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.081408] env[61986]: DEBUG nova.network.neutron [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.313666] env[61986]: DEBUG nova.compute.utils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.317964] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.318173] env[61986]: DEBUG nova.network.neutron [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.350525] env[61986]: DEBUG nova.network.neutron [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Updated VIF entry in instance network info cache for port 1234061c-4805-4e9a-9993-3e09549f5569. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.350882] env[61986]: DEBUG nova.network.neutron [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Updating instance_info_cache with network_info: [{"id": "1234061c-4805-4e9a-9993-3e09549f5569", "address": "fa:16:3e:3c:45:e9", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1234061c-48", "ovs_interfaceid": "1234061c-4805-4e9a-9993-3e09549f5569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.384943] env[61986]: DEBUG nova.policy [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9bf0c90750f142689cc84162c73facf4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '452cd8b9693849218d720a05cf3248c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 873.407110] env[61986]: INFO nova.compute.manager [None req-1431f492-fbbd-444e-96f7-ed9563c20ab2 tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance to original state: 'active' [ 873.431517] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159814, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.483181] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159813, 'name': CreateVM_Task, 'duration_secs': 0.538972} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.483396] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 873.484013] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.484199] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.484598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.484861] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-991cbf46-d8a3-48ee-98a3-81d17184ef7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.490405] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 873.490405] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5262ac9a-02de-a9d3-99a4-42a83b22f3da" [ 873.490405] env[61986]: _type = "Task" [ 873.490405] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.503854] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5262ac9a-02de-a9d3-99a4-42a83b22f3da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.534150] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.628486] env[61986]: DEBUG nova.network.neutron [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 873.799228] env[61986]: DEBUG nova.network.neutron [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Updating instance_info_cache with network_info: [{"id": "7d08d53f-b8f7-4397-acce-ff7cdeeb15c0", "address": "fa:16:3e:a9:c0:87", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d08d53f-b8", "ovs_interfaceid": "7d08d53f-b8f7-4397-acce-ff7cdeeb15c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.820041] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.833983] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573b7a22-e992-4a1c-b1ce-2e705755e100 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.842654] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76187699-b105-4121-9ed7-e9142a658af2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.877627] env[61986]: DEBUG oslo_concurrency.lockutils [req-ffbed9d4-06da-4396-8875-7f436dcdb2ce req-12d2f68c-d960-4a58-a8c0-f1e106465857 service nova] Releasing lock "refresh_cache-b9706e51-0f74-4dbc-aab1-ea640ac78c62" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.879823] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f385b123-173f-48c5-b74f-b9b9f812598c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.888674] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1af22ca-8e08-4b5d-b3e9-c41381e48ba7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.903997] env[61986]: DEBUG nova.compute.provider_tree [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.939018] env[61986]: DEBUG oslo_vmware.api [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159814, 'name': PowerOnVM_Task, 'duration_secs': 0.784053} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.939018] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.939018] env[61986]: INFO nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Took 10.76 seconds to spawn the instance on the hypervisor. [ 873.939018] env[61986]: DEBUG nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.939018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f23fdea-5c5d-4313-8ec7-11e9e41ff917 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.944929] env[61986]: DEBUG nova.network.neutron [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Successfully created port: 494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.002074] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5262ac9a-02de-a9d3-99a4-42a83b22f3da, 'name': SearchDatastore_Task, 'duration_secs': 0.020814} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.002405] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.002649] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.002879] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.003037] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.003225] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.003496] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-298d06a0-03b6-4194-b8fa-0de1dc0fcf95 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.015233] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.015455] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 874.016361] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cca335bd-2847-4e5e-a3a2-b0e9787ba888 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.027319] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 874.027319] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f6ca4a-2efd-3784-ac71-218a149c7b89" [ 874.027319] env[61986]: _type = "Task" [ 874.027319] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.035217] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.041561] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f6ca4a-2efd-3784-ac71-218a149c7b89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.305132] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Releasing lock "refresh_cache-4f13b629-e2a7-4668-9d77-eb638078e246" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.305518] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Instance network_info: |[{"id": "7d08d53f-b8f7-4397-acce-ff7cdeeb15c0", "address": "fa:16:3e:a9:c0:87", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d08d53f-b8", "ovs_interfaceid": "7d08d53f-b8f7-4397-acce-ff7cdeeb15c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 874.306022] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:c0:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d08d53f-b8f7-4397-acce-ff7cdeeb15c0', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.315300] env[61986]: DEBUG oslo.service.loopingcall [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.315779] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 874.315911] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf7718c0-b456-4292-ada3-54462ed4d0a5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.336265] env[61986]: INFO nova.virt.block_device [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Booting with volume 6d6d6476-0f93-4404-b58d-3092cea80b1b at /dev/sda [ 874.356884] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.356884] env[61986]: value = "task-1159816" [ 874.356884] env[61986]: _type = "Task" [ 874.356884] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.366885] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159816, 'name': CreateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.393975] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f49a0b7a-0ce9-4ecb-b1b2-a5fa1d150b45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.405783] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc889d3-70f5-4d48-9148-3b17d1c30449 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.418829] env[61986]: DEBUG nova.scheduler.client.report [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.466234] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35821ecc-150b-4821-bdf7-b3be23405214 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.469309] env[61986]: INFO nova.compute.manager [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Took 44.11 seconds to build instance. [ 874.480832] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c2c112-6bfd-47ca-87a1-641461f8cb37 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.517673] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb06ffa0-5934-45a6-b19e-5c28cab3ac32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.529117] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0766fc2b-0f65-4a0f-bdcf-84764cd97fb2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.541249] env[61986]: DEBUG oslo_vmware.api [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159801, 'name': ReconfigVM_Task, 'duration_secs': 6.391983} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.544975] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.545276] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Reconfigured VM to detach interface {{(pid=61986) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 874.547777] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f6ca4a-2efd-3784-ac71-218a149c7b89, 'name': SearchDatastore_Task, 'duration_secs': 0.02054} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.549118] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18794cff-879c-4833-aef6-4621f94257c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.561698] env[61986]: DEBUG nova.virt.block_device [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updating existing volume attachment record: 4628806d-d4ed-4cc8-b048-a59b6deb24ea {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 874.566173] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 874.566173] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522763a5-3aa4-c83c-8dff-d20d1729b3e4" [ 874.566173] env[61986]: _type = "Task" [ 874.566173] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.576271] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522763a5-3aa4-c83c-8dff-d20d1729b3e4, 'name': SearchDatastore_Task, 'duration_secs': 0.01319} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.576589] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.576812] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] b9706e51-0f74-4dbc-aab1-ea640ac78c62/b9706e51-0f74-4dbc-aab1-ea640ac78c62.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.578906] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7688cf49-22f0-4de3-a2b1-d81c22124ad4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.590257] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 874.590257] env[61986]: value = "task-1159817" [ 874.590257] env[61986]: _type = "Task" [ 874.590257] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.603648] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159817, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.864741] env[61986]: DEBUG nova.compute.manager [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Received event network-changed-7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.865030] env[61986]: DEBUG nova.compute.manager [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Refreshing instance network info cache due to event network-changed-7d08d53f-b8f7-4397-acce-ff7cdeeb15c0. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 874.865220] env[61986]: DEBUG oslo_concurrency.lockutils [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] Acquiring lock "refresh_cache-4f13b629-e2a7-4668-9d77-eb638078e246" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.865393] env[61986]: DEBUG oslo_concurrency.lockutils [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] Acquired lock "refresh_cache-4f13b629-e2a7-4668-9d77-eb638078e246" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.865555] env[61986]: DEBUG nova.network.neutron [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Refreshing network info cache for port 7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 874.874379] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159816, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.924631] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.114s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.928133] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.999s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.930982] env[61986]: INFO nova.compute.claims [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.956356] env[61986]: INFO nova.scheduler.client.report [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocations for instance 6f20dd8e-9289-458d-bed6-f2ef9daaa917 [ 874.971506] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b70ffc49-4360-498b-92ed-773a2c804bfc tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.965s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.102616] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159817, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.372252] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159816, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.467095] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8c6a6697-003e-4bd9-9bb1-31e06768ec23 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "6f20dd8e-9289-458d-bed6-f2ef9daaa917" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.532s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.475863] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 875.609930] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159817, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738054} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.610939] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] b9706e51-0f74-4dbc-aab1-ea640ac78c62/b9706e51-0f74-4dbc-aab1-ea640ac78c62.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 875.610939] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.610939] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-001b125c-a76a-4735-88e1-a8782ba46dbb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.620629] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 875.620629] env[61986]: value = "task-1159818" [ 875.620629] env[61986]: _type = "Task" [ 875.620629] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.625393] env[61986]: DEBUG nova.network.neutron [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Updated VIF entry in instance network info cache for port 7d08d53f-b8f7-4397-acce-ff7cdeeb15c0. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 875.625942] env[61986]: DEBUG nova.network.neutron [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Updating instance_info_cache with network_info: [{"id": "7d08d53f-b8f7-4397-acce-ff7cdeeb15c0", "address": "fa:16:3e:a9:c0:87", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d08d53f-b8", "ovs_interfaceid": "7d08d53f-b8f7-4397-acce-ff7cdeeb15c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.641174] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159818, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.822086] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "40de5f22-0de2-466a-91ab-dcb6ec586dad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.822179] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.822519] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.822636] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.822872] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.825428] env[61986]: INFO nova.compute.manager [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Terminating instance [ 875.827103] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.827359] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.827604] env[61986]: DEBUG nova.network.neutron [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.829292] env[61986]: DEBUG nova.compute.manager [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.829448] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 875.830589] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2453e06-2c5b-41a0-89a7-051558212ef1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.839705] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 875.840745] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d0a19fc-e1b6-40b9-ba9e-1f4a4ba3f656 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.850792] env[61986]: DEBUG oslo_vmware.api [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 875.850792] env[61986]: value = "task-1159820" [ 875.850792] env[61986]: _type = "Task" [ 875.850792] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.863272] env[61986]: DEBUG oslo_vmware.api [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.873662] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159816, 'name': CreateVM_Task, 'duration_secs': 1.488955} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.873967] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 875.874834] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.875095] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.875451] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.875767] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66b9fc39-b23c-4744-a6a3-5ec9231a41d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.881317] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 875.881317] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52205266-a168-4a52-0939-c96c8e570045" [ 875.881317] env[61986]: _type = "Task" [ 875.881317] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.889688] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52205266-a168-4a52-0939-c96c8e570045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.993893] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.056484] env[61986]: DEBUG nova.network.neutron [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Successfully updated port: 494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.127796] env[61986]: DEBUG nova.compute.manager [req-8e3920ae-7874-4f76-aa0e-cf559b38cf24 req-470a631c-8a93-49b8-bfa4-6ccf1bfff323 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Received event network-vif-plugged-494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.128058] env[61986]: DEBUG oslo_concurrency.lockutils [req-8e3920ae-7874-4f76-aa0e-cf559b38cf24 req-470a631c-8a93-49b8-bfa4-6ccf1bfff323 service nova] Acquiring lock "98337bb6-9502-4d4c-af00-028659b246bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.128477] env[61986]: DEBUG oslo_concurrency.lockutils [req-8e3920ae-7874-4f76-aa0e-cf559b38cf24 req-470a631c-8a93-49b8-bfa4-6ccf1bfff323 service nova] Lock "98337bb6-9502-4d4c-af00-028659b246bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.128718] env[61986]: DEBUG oslo_concurrency.lockutils [req-8e3920ae-7874-4f76-aa0e-cf559b38cf24 req-470a631c-8a93-49b8-bfa4-6ccf1bfff323 service nova] Lock "98337bb6-9502-4d4c-af00-028659b246bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.128904] env[61986]: DEBUG nova.compute.manager [req-8e3920ae-7874-4f76-aa0e-cf559b38cf24 req-470a631c-8a93-49b8-bfa4-6ccf1bfff323 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] No waiting events found dispatching network-vif-plugged-494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 876.129115] env[61986]: WARNING nova.compute.manager [req-8e3920ae-7874-4f76-aa0e-cf559b38cf24 req-470a631c-8a93-49b8-bfa4-6ccf1bfff323 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Received unexpected event network-vif-plugged-494be155-c4b7-446d-b2ca-d49f22c7cc85 for instance with vm_state building and task_state block_device_mapping. [ 876.134939] env[61986]: DEBUG oslo_concurrency.lockutils [req-7d63e4a7-8e4c-4487-b93e-e7a7a9304e61 req-fae8d911-d098-4c1e-bebf-8a32b1e817ba service nova] Releasing lock "refresh_cache-4f13b629-e2a7-4668-9d77-eb638078e246" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.139291] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159818, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125801} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.139554] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.140476] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc458ba-67f2-4a99-b75b-6a7177da2310 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.172216] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] b9706e51-0f74-4dbc-aab1-ea640ac78c62/b9706e51-0f74-4dbc-aab1-ea640ac78c62.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.172216] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4e0e93e-a40f-4293-8e78-4b1ff2796164 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.195486] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 876.195486] env[61986]: value = "task-1159821" [ 876.195486] env[61986]: _type = "Task" [ 876.195486] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.214206] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159821, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.252989] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.253370] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "977799b4-2793-4513-9447-483146fc7ac4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.253664] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "977799b4-2793-4513-9447-483146fc7ac4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.253905] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.254312] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "977799b4-2793-4513-9447-483146fc7ac4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.256695] env[61986]: INFO nova.compute.manager [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Terminating instance [ 876.259177] env[61986]: DEBUG nova.compute.manager [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.259401] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 876.262861] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2376257-aa2e-4766-ae12-0e85817f705d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.271507] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 876.271885] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09326d5c-4af1-4e11-b612-506f2b6a58e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.280937] env[61986]: DEBUG oslo_vmware.api [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 876.280937] env[61986]: value = "task-1159822" [ 876.280937] env[61986]: _type = "Task" [ 876.280937] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.299171] env[61986]: DEBUG oslo_vmware.api [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.365560] env[61986]: DEBUG oslo_vmware.api [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159820, 'name': PowerOffVM_Task, 'duration_secs': 0.392233} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.366265] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 876.366265] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 876.366825] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a62913f2-c8ad-475f-9f3d-5875e5903748 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.398978] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52205266-a168-4a52-0939-c96c8e570045, 'name': SearchDatastore_Task, 'duration_secs': 0.03534} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.399243] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.399462] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.399717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.399875] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.400072] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.400408] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd370ccd-173d-43d3-a63f-c9ad4350ed00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.413751] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.413908] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.414799] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8feb6e9c-4ecf-4654-ae48-52de58b0851d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.423922] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 876.423922] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52125b34-ed94-e4dd-6302-a60e241ba5a6" [ 876.423922] env[61986]: _type = "Task" [ 876.423922] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.436493] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52125b34-ed94-e4dd-6302-a60e241ba5a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.439529] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 876.439715] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 876.439904] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Deleting the datastore file [datastore1] 40de5f22-0de2-466a-91ab-dcb6ec586dad {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.440188] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5abc8bdf-eb86-44c1-9d71-d394a7fc6790 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.447484] env[61986]: DEBUG oslo_vmware.api [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for the task: (returnval){ [ 876.447484] env[61986]: value = "task-1159824" [ 876.447484] env[61986]: _type = "Task" [ 876.447484] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.456581] env[61986]: DEBUG oslo_vmware.api [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.472640] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736dd827-4b9b-47f2-a5bb-4f536be12d6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.487589] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9089fe-689e-427b-88b0-c1ae6fb3b328 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.521607] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36916fd0-3ae9-43ab-b12b-6549c40360b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.532555] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63a88a8-0f4b-4451-9467-725f16e182bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.551404] env[61986]: DEBUG nova.compute.provider_tree [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.565561] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.565561] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquired lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.565561] env[61986]: DEBUG nova.network.neutron [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.707648] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.708230] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.708482] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.708660] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.708845] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.708995] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.709167] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.709407] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.709577] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.709748] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.710143] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.710255] env[61986]: DEBUG nova.virt.hardware [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.710595] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159821, 'name': ReconfigVM_Task, 'duration_secs': 0.422849} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.711361] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae68307c-4e06-4bcc-b8ef-c1725c64ef71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.714040] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Reconfigured VM instance instance-0000003e to attach disk [datastore2] b9706e51-0f74-4dbc-aab1-ea640ac78c62/b9706e51-0f74-4dbc-aab1-ea640ac78c62.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.715000] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cded360a-bf11-49b4-b3ef-26daefe6a2b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.722883] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454e0f8b-8ed0-4a5d-b2d9-a5de8327de8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.729716] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 876.729716] env[61986]: value = "task-1159825" [ 876.729716] env[61986]: _type = "Task" [ 876.729716] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.747717] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159825, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.791856] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.792141] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.792361] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.792548] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.792723] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.794505] env[61986]: DEBUG oslo_vmware.api [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159822, 'name': PowerOffVM_Task, 'duration_secs': 0.317772} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.794766] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 876.794955] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 876.795245] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4b586f1-1080-4b0f-9590-397a1e3ad6ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.798315] env[61986]: INFO nova.compute.manager [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Terminating instance [ 876.799957] env[61986]: DEBUG nova.compute.manager [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.800169] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 876.800962] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f252c69d-258f-4571-b36b-04690c2cbd1a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.809481] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 876.809759] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc9064ed-4eaa-47f5-acb2-cd1fa4763ccd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.816775] env[61986]: DEBUG oslo_vmware.api [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 876.816775] env[61986]: value = "task-1159827" [ 876.816775] env[61986]: _type = "Task" [ 876.816775] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.829231] env[61986]: DEBUG oslo_vmware.api [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.837200] env[61986]: INFO nova.network.neutron [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Port b2150478-f9cd-4dba-94fe-205081032d36 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 876.837829] env[61986]: INFO nova.network.neutron [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Port 5a6caae3-e994-43ff-b2b4-5e3f89aea659 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 876.837829] env[61986]: DEBUG nova.network.neutron [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [{"id": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "address": "fa:16:3e:ae:b8:61", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c8749d8-2c", "ovs_interfaceid": "4c8749d8-2c4f-455b-9e9f-c4dc823cc295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.891557] env[61986]: DEBUG oslo_concurrency.lockutils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.891805] env[61986]: DEBUG oslo_concurrency.lockutils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.935579] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52125b34-ed94-e4dd-6302-a60e241ba5a6, 'name': SearchDatastore_Task, 'duration_secs': 0.011605} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.937169] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-488d8278-fccb-412b-92ed-a757707339bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.944155] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 876.944155] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b815c6-f5ae-dc79-4b1f-28bfc5ffdf69" [ 876.944155] env[61986]: _type = "Task" [ 876.944155] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.948921] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 876.949217] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 876.949431] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleting the datastore file [datastore2] 977799b4-2793-4513-9447-483146fc7ac4 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.953612] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1034217b-8847-4b9c-8fa1-378546e48d8c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.958162] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b815c6-f5ae-dc79-4b1f-28bfc5ffdf69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.963982] env[61986]: DEBUG oslo_vmware.api [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Task: {'id': task-1159824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.313361} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.965510] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.965807] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 876.966133] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 876.966368] env[61986]: INFO nova.compute.manager [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Took 1.14 seconds to destroy the instance on the hypervisor. [ 876.966618] env[61986]: DEBUG oslo.service.loopingcall [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.966883] env[61986]: DEBUG oslo_vmware.api [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 876.966883] env[61986]: value = "task-1159828" [ 876.966883] env[61986]: _type = "Task" [ 876.966883] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.967088] env[61986]: DEBUG nova.compute.manager [-] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.967187] env[61986]: DEBUG nova.network.neutron [-] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 876.977654] env[61986]: DEBUG oslo_vmware.api [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.055073] env[61986]: DEBUG nova.scheduler.client.report [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.109983] env[61986]: DEBUG nova.network.neutron [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.241896] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159825, 'name': Rename_Task, 'duration_secs': 0.182514} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.242229] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.242486] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a692ebd4-5fd7-4de9-8b2a-3e37a612aa12 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.252272] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 877.252272] env[61986]: value = "task-1159829" [ 877.252272] env[61986]: _type = "Task" [ 877.252272] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.256632] env[61986]: DEBUG nova.network.neutron [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updating instance_info_cache with network_info: [{"id": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "address": "fa:16:3e:85:11:df", "network": {"id": "10ba456d-8a3c-4689-a339-7b9e216a2680", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1150131841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "452cd8b9693849218d720a05cf3248c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494be155-c4", "ovs_interfaceid": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.263238] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.331077] env[61986]: DEBUG oslo_vmware.api [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159827, 'name': PowerOffVM_Task, 'duration_secs': 0.266196} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.333078] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 877.333473] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 877.334403] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-770b2176-e757-4d16-82d9-164c0ccdf0c1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.341035] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-977799b4-2793-4513-9447-483146fc7ac4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.395671] env[61986]: DEBUG nova.compute.utils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.426998] env[61986]: DEBUG nova.compute.manager [req-263d121d-f107-4e88-994f-3789247eb739 req-c1abf3f0-607c-4b54-960e-80a460e52b52 service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-vif-deleted-b2150478-f9cd-4dba-94fe-205081032d36 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.426998] env[61986]: DEBUG nova.compute.manager [req-263d121d-f107-4e88-994f-3789247eb739 req-c1abf3f0-607c-4b54-960e-80a460e52b52 service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-vif-deleted-5a6caae3-e994-43ff-b2b4-5e3f89aea659 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.438471] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 877.438703] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 877.438894] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleting the datastore file [datastore1] 46b18b6c-6e40-45b7-9d3f-6177b08a52a5 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.439281] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50a5b7b9-da6b-49a4-88ae-6cd83f924d50 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.449656] env[61986]: DEBUG oslo_vmware.api [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 877.449656] env[61986]: value = "task-1159831" [ 877.449656] env[61986]: _type = "Task" [ 877.449656] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.457094] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b815c6-f5ae-dc79-4b1f-28bfc5ffdf69, 'name': SearchDatastore_Task, 'duration_secs': 0.011554} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.457786] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.458063] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 4f13b629-e2a7-4668-9d77-eb638078e246/4f13b629-e2a7-4668-9d77-eb638078e246.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.458372] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0f47575-2fc6-49a8-a2e7-08fcfd03db99 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.463529] env[61986]: DEBUG oslo_vmware.api [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159831, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.469986] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 877.469986] env[61986]: value = "task-1159832" [ 877.469986] env[61986]: _type = "Task" [ 877.469986] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.481567] env[61986]: DEBUG oslo_vmware.api [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1159828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174207} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.484808] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.485013] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 877.485224] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 877.485408] env[61986]: INFO nova.compute.manager [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Took 1.23 seconds to destroy the instance on the hypervisor. [ 877.485673] env[61986]: DEBUG oslo.service.loopingcall [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.485875] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159832, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.486129] env[61986]: DEBUG nova.compute.manager [-] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.486244] env[61986]: DEBUG nova.network.neutron [-] [instance: 977799b4-2793-4513-9447-483146fc7ac4] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 877.560735] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.633s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.561072] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.566863] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 25.458s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.567082] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.567274] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 877.567638] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.376s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.567898] env[61986]: DEBUG nova.objects.instance [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lazy-loading 'resources' on Instance uuid 182479b8-f72f-4395-99a4-af0d6f91f7d4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.570415] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42907bb-d039-4f7f-80d2-552bf9fe9599 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.581065] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eca0812-653e-429e-a675-5fa84041749a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.587870] env[61986]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 5a6caae3-e994-43ff-b2b4-5e3f89aea659 could not be found.", "detail": ""}} {{(pid=61986) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 877.588244] env[61986]: DEBUG nova.network.neutron [-] Unable to show port 5a6caae3-e994-43ff-b2b4-5e3f89aea659 as it no longer exists. {{(pid=61986) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 877.600601] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d1841f-053a-4cda-8161-ae2c74d57428 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.608943] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46e74d0-cab4-49e9-8900-a04a2d40d687 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.645141] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179676MB free_disk=162GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 877.645486] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.763346] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Releasing lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.765826] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance network_info: |[{"id": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "address": "fa:16:3e:85:11:df", "network": {"id": "10ba456d-8a3c-4689-a339-7b9e216a2680", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1150131841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "452cd8b9693849218d720a05cf3248c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494be155-c4", "ovs_interfaceid": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.765826] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159829, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.765826] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:11:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '494be155-c4b7-446d-b2ca-d49f22c7cc85', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.772573] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Creating folder: Project (452cd8b9693849218d720a05cf3248c1). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 877.773316] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b868c72-2a0a-4eb1-83ba-358ff43a52f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.790791] env[61986]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 877.790977] env[61986]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61986) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 877.791437] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Folder already exists: Project (452cd8b9693849218d720a05cf3248c1). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.791653] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Creating folder: Instances. Parent ref: group-v252287. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 877.791920] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3e98207-6c17-45d6-a0bb-4b3b075e229f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.805375] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Created folder: Instances in parent group-v252287. [ 877.806615] env[61986]: DEBUG oslo.service.loopingcall [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.806959] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 877.807457] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9406b1b-9573-454c-a6b2-f9f5b328f2a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.834648] env[61986]: DEBUG nova.network.neutron [-] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.839718] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.839718] env[61986]: value = "task-1159836" [ 877.839718] env[61986]: _type = "Task" [ 877.839718] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.846128] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2d500319-720a-4bf1-93ca-78b6e8236972 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-977799b4-2793-4513-9447-483146fc7ac4-b2150478-f9cd-4dba-94fe-205081032d36" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.424s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.852027] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159836, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.899389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.962200] env[61986]: DEBUG oslo_vmware.api [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159831, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272928} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.962468] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.962653] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 877.962861] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 877.963147] env[61986]: INFO nova.compute.manager [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 877.963277] env[61986]: DEBUG oslo.service.loopingcall [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.963487] env[61986]: DEBUG nova.compute.manager [-] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.963578] env[61986]: DEBUG nova.network.neutron [-] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 877.985584] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159832, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.074693] env[61986]: DEBUG nova.compute.utils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.079400] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 878.079400] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 878.176034] env[61986]: DEBUG nova.policy [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb4b3a6fb1c7426796f38978b04940d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdab411a34424bc185f8a5981809437c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 878.266745] env[61986]: DEBUG oslo_vmware.api [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159829, 'name': PowerOnVM_Task, 'duration_secs': 0.737375} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.271921] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.271921] env[61986]: INFO nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Took 8.16 seconds to spawn the instance on the hypervisor. [ 878.271921] env[61986]: DEBUG nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.273165] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6052a56b-c3aa-42b6-8252-61dbf81a5ec0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.338994] env[61986]: DEBUG nova.compute.manager [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.339728] env[61986]: INFO nova.compute.manager [-] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Took 1.37 seconds to deallocate network for instance. [ 878.341574] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57148ed-412a-4853-9b46-ca7f3e75bf4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.372623] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159836, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.489471] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159832, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685579} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.490079] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 4f13b629-e2a7-4668-9d77-eb638078e246/4f13b629-e2a7-4668-9d77-eb638078e246.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.490524] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.493460] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a658a7db-8e76-4b3c-b464-974c8ea061f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.501577] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 878.501577] env[61986]: value = "task-1159837" [ 878.501577] env[61986]: _type = "Task" [ 878.501577] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.516302] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159837, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.575984] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Successfully created port: 487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.583028] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.704766] env[61986]: DEBUG nova.network.neutron [-] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.707840] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c43891-8540-4264-9886-df10889f1c59 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.714296] env[61986]: DEBUG nova.compute.manager [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Received event network-changed-494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.715817] env[61986]: DEBUG nova.compute.manager [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Refreshing instance network info cache due to event network-changed-494be155-c4b7-446d-b2ca-d49f22c7cc85. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 878.715817] env[61986]: DEBUG oslo_concurrency.lockutils [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] Acquiring lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.715817] env[61986]: DEBUG oslo_concurrency.lockutils [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] Acquired lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.715817] env[61986]: DEBUG nova.network.neutron [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Refreshing network info cache for port 494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.722493] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b65a4a7-210d-448a-96a9-3ef8ba746a35 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.758807] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc37e0a-749b-4513-987a-83df6193dd6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.769611] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a4dd3f-0c10-475c-bb05-5d9475fba932 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.784105] env[61986]: DEBUG nova.compute.provider_tree [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.800651] env[61986]: INFO nova.compute.manager [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Took 39.70 seconds to build instance. [ 878.850595] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159836, 'name': CreateVM_Task, 'duration_secs': 0.598509} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.850846] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 878.852031] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252302', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'name': 'volume-6d6d6476-0f93-4404-b58d-3092cea80b1b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98337bb6-9502-4d4c-af00-028659b246bf', 'attached_at': '', 'detached_at': '', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'serial': '6d6d6476-0f93-4404-b58d-3092cea80b1b'}, 'guest_format': None, 'attachment_id': '4628806d-d4ed-4cc8-b048-a59b6deb24ea', 'boot_index': 0, 'disk_bus': None, 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61986) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 878.852230] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Root volume attach. Driver type: vmdk {{(pid=61986) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 878.853097] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384371ca-767d-4e66-bf25-48de8b57570b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.862618] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.862813] env[61986]: DEBUG nova.network.neutron [-] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.868019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4782f16c-c76e-45d7-98c9-4f23c325b3b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.875340] env[61986]: INFO nova.compute.manager [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] instance snapshotting [ 878.875340] env[61986]: DEBUG nova.objects.instance [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'flavor' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.878406] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c935f2-ecf4-4f23-830e-2f5407003fc3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.889229] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-95b2aaa4-a561-43cb-8ebc-4a93630183d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.897893] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 878.897893] env[61986]: value = "task-1159838" [ 878.897893] env[61986]: _type = "Task" [ 878.897893] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.906961] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159838, 'name': RelocateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.002674] env[61986]: DEBUG oslo_concurrency.lockutils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.003040] env[61986]: DEBUG oslo_concurrency.lockutils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.003302] env[61986]: INFO nova.compute.manager [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Attaching volume 853747f9-2ca2-43a2-acfb-5d955470ace1 to /dev/sdb [ 879.016850] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159837, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075603} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.016850] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.017328] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ece932-d5f5-42a7-80fa-4e06853fee5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.041402] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 4f13b629-e2a7-4668-9d77-eb638078e246/4f13b629-e2a7-4668-9d77-eb638078e246.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.042558] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-257b9421-a5a3-491c-ab25-431acbc76ef3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.058211] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108ddd1a-b4a2-44c4-8b98-a6fa38dc4e42 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.066233] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d40e99-7968-4eae-9965-5737b780f15a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.071151] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 879.071151] env[61986]: value = "task-1159839" [ 879.071151] env[61986]: _type = "Task" [ 879.071151] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.083692] env[61986]: DEBUG nova.virt.block_device [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updating existing volume attachment record: ebb17a93-1b83-4ffa-b78c-b206da95d268 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 879.089349] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159839, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.212363] env[61986]: INFO nova.compute.manager [-] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Took 1.73 seconds to deallocate network for instance. [ 879.288775] env[61986]: DEBUG nova.scheduler.client.report [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.305086] env[61986]: DEBUG oslo_concurrency.lockutils [None req-13937bb4-44f6-4536-a5d3-200f7eda958a tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.252s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.366439] env[61986]: INFO nova.compute.manager [-] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Took 1.40 seconds to deallocate network for instance. [ 879.387302] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204bb907-a453-4702-9721-8f9106f499d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.414529] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8bdfbe-e699-431f-845e-d81803a2cc92 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.422866] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159838, 'name': RelocateVM_Task, 'duration_secs': 0.03039} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.425023] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 879.425402] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252302', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'name': 'volume-6d6d6476-0f93-4404-b58d-3092cea80b1b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98337bb6-9502-4d4c-af00-028659b246bf', 'attached_at': '', 'detached_at': '', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'serial': '6d6d6476-0f93-4404-b58d-3092cea80b1b'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 879.428857] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a15a4b-2b4d-4d86-a06e-ca2ec7d093f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.449304] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de008ca6-6e38-470f-8928-085dac243407 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.475981] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] volume-6d6d6476-0f93-4404-b58d-3092cea80b1b/volume-6d6d6476-0f93-4404-b58d-3092cea80b1b.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.478018] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d949221-6b3b-4381-a292-4941af8f24eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.493277] env[61986]: DEBUG nova.compute.manager [req-76c50980-4979-4219-a0f1-69f84e1d2d35 req-af9e2620-9c0e-4d93-a6e4-f1c695690ba8 service nova] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Received event network-vif-deleted-8b44e4ce-d172-4864-90e3-3c6cf9486bfb {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.500731] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 879.500731] env[61986]: value = "task-1159843" [ 879.500731] env[61986]: _type = "Task" [ 879.500731] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.513183] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159843, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.544692] env[61986]: DEBUG nova.network.neutron [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updated VIF entry in instance network info cache for port 494be155-c4b7-446d-b2ca-d49f22c7cc85. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.545304] env[61986]: DEBUG nova.network.neutron [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updating instance_info_cache with network_info: [{"id": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "address": "fa:16:3e:85:11:df", "network": {"id": "10ba456d-8a3c-4689-a339-7b9e216a2680", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1150131841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "452cd8b9693849218d720a05cf3248c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494be155-c4", "ovs_interfaceid": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.581926] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.601942] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.631752] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.632032] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.632201] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.632507] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.632557] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.632680] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.632889] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.633115] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.633244] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.633439] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.633617] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.634534] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525bbac3-90ee-42c0-b514-adb669f08644 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.643748] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d0b3b6-8327-4a92-b2ee-0ad17a529f67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.721924] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.760851] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.761122] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.795431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.227s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.797918] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.898s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.799493] env[61986]: INFO nova.compute.claims [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.810024] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.820246] env[61986]: INFO nova.scheduler.client.report [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Deleted allocations for instance 182479b8-f72f-4395-99a4-af0d6f91f7d4 [ 879.879769] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.934577] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 879.934577] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-12c26710-4f92-44e6-a999-74cd602d29ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.948251] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 879.948251] env[61986]: value = "task-1159845" [ 879.948251] env[61986]: _type = "Task" [ 879.948251] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.959725] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159845, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.018466] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159843, 'name': ReconfigVM_Task, 'duration_secs': 0.290094} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.019020] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Reconfigured VM instance instance-00000040 to attach disk [datastore2] volume-6d6d6476-0f93-4404-b58d-3092cea80b1b/volume-6d6d6476-0f93-4404-b58d-3092cea80b1b.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.024283] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe292d82-034e-49b5-ba51-c52850cf170a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.042115] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 880.042115] env[61986]: value = "task-1159846" [ 880.042115] env[61986]: _type = "Task" [ 880.042115] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.051939] env[61986]: DEBUG oslo_concurrency.lockutils [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] Releasing lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.051939] env[61986]: DEBUG nova.compute.manager [req-6b8279b3-84f7-4d18-aa82-9ac508c9948c req-9e9e4ed8-5ac3-4967-ab61-88e8423d659a service nova] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Received event network-vif-deleted-dd125b45-dffb-4a25-af78-4a9326408be0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.052132] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.082462] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159839, 'name': ReconfigVM_Task, 'duration_secs': 0.625546} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.082462] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 4f13b629-e2a7-4668-9d77-eb638078e246/4f13b629-e2a7-4668-9d77-eb638078e246.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.085032] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61f3393d-1679-4ee9-9aef-7124670dca34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.091164] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 880.091164] env[61986]: value = "task-1159847" [ 880.091164] env[61986]: _type = "Task" [ 880.091164] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.100120] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159847, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.138620] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Successfully updated port: 487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.331240] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.332166] env[61986]: DEBUG oslo_concurrency.lockutils [None req-67e50ece-ca4e-402a-95cc-63ea7e75ff26 tempest-ServerGroupTestJSON-45905315 tempest-ServerGroupTestJSON-45905315-project-member] Lock "182479b8-f72f-4395-99a4-af0d6f91f7d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.112s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.458827] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159845, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.552806] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159846, 'name': ReconfigVM_Task, 'duration_secs': 0.134403} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.553127] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252302', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'name': 'volume-6d6d6476-0f93-4404-b58d-3092cea80b1b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98337bb6-9502-4d4c-af00-028659b246bf', 'attached_at': '', 'detached_at': '', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'serial': '6d6d6476-0f93-4404-b58d-3092cea80b1b'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 880.553677] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1040b7fe-52e2-449c-b0e1-db3b6f0e82fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.561277] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 880.561277] env[61986]: value = "task-1159848" [ 880.561277] env[61986]: _type = "Task" [ 880.561277] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.572471] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159848, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.602026] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159847, 'name': Rename_Task, 'duration_secs': 0.175056} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.602170] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.602505] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7db20a2-bbde-4705-817f-73a60e6cf1bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.610918] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 880.610918] env[61986]: value = "task-1159849" [ 880.610918] env[61986]: _type = "Task" [ 880.610918] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.619588] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.642563] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "refresh_cache-9b539a94-6c39-46f8-b194-27047245d1f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.642820] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "refresh_cache-9b539a94-6c39-46f8-b194-27047245d1f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.643069] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.761300] env[61986]: DEBUG nova.compute.manager [req-4fe5f51b-181a-4851-87fb-d908bcb3a168 req-f546702e-2d4d-40a5-aa55-52d176b2de4e service nova] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Received event network-vif-deleted-4c8749d8-2c4f-455b-9e9f-c4dc823cc295 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.962740] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159845, 'name': CreateSnapshot_Task, 'duration_secs': 0.917344} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.963108] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 880.963746] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d187d9-d8f8-46ab-90a4-ce2d02dda0fc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.077946] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159848, 'name': Rename_Task, 'duration_secs': 0.133769} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.077946] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 881.077946] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-439f6e37-cf57-47dd-8805-81d8c4f8c3d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.084041] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 881.084041] env[61986]: value = "task-1159850" [ 881.084041] env[61986]: _type = "Task" [ 881.084041] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.094947] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.123645] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159849, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.198145] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 881.255738] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994f5334-fa97-4833-b532-d8fd5298c68b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.269054] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56744ffe-8531-495e-b78d-a9f7c47b97b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.306921] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbcb962-78b0-4c12-8e22-9d4132023888 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.316639] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bf8b65-56c8-4e8c-a963-5c22d0e42b78 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.332920] env[61986]: DEBUG nova.compute.provider_tree [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.442168] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Updating instance_info_cache with network_info: [{"id": "487bb20b-a111-4582-922e-c52404bb758a", "address": "fa:16:3e:85:8f:72", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487bb20b-a1", "ovs_interfaceid": "487bb20b-a111-4582-922e-c52404bb758a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.486289] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 881.486740] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cb028b77-9999-4417-bba9-5d66bd5cbbc8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.496490] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 881.496490] env[61986]: value = "task-1159851" [ 881.496490] env[61986]: _type = "Task" [ 881.496490] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.506480] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159851, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.508714] env[61986]: DEBUG nova.compute.manager [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Received event network-vif-plugged-487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.509062] env[61986]: DEBUG oslo_concurrency.lockutils [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] Acquiring lock "9b539a94-6c39-46f8-b194-27047245d1f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.509370] env[61986]: DEBUG oslo_concurrency.lockutils [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] Lock "9b539a94-6c39-46f8-b194-27047245d1f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.509884] env[61986]: DEBUG oslo_concurrency.lockutils [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] Lock "9b539a94-6c39-46f8-b194-27047245d1f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.509884] env[61986]: DEBUG nova.compute.manager [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] No waiting events found dispatching network-vif-plugged-487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.510039] env[61986]: WARNING nova.compute.manager [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Received unexpected event network-vif-plugged-487bb20b-a111-4582-922e-c52404bb758a for instance with vm_state building and task_state spawning. [ 881.510143] env[61986]: DEBUG nova.compute.manager [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Received event network-changed-487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.510602] env[61986]: DEBUG nova.compute.manager [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Refreshing instance network info cache due to event network-changed-487bb20b-a111-4582-922e-c52404bb758a. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 881.510602] env[61986]: DEBUG oslo_concurrency.lockutils [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] Acquiring lock "refresh_cache-9b539a94-6c39-46f8-b194-27047245d1f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.595300] env[61986]: DEBUG oslo_vmware.api [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159850, 'name': PowerOnVM_Task, 'duration_secs': 0.49922} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.595660] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.595887] env[61986]: INFO nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Took 4.89 seconds to spawn the instance on the hypervisor. [ 881.596085] env[61986]: DEBUG nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.597955] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37704337-0072-41e5-a30a-afe2855cbd4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.625411] env[61986]: DEBUG oslo_vmware.api [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159849, 'name': PowerOnVM_Task, 'duration_secs': 0.823132} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.625689] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.626244] env[61986]: INFO nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Took 9.43 seconds to spawn the instance on the hypervisor. [ 881.626244] env[61986]: DEBUG nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.627030] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd645823-5ed4-419a-8f08-dc1c777546f8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.836087] env[61986]: DEBUG nova.scheduler.client.report [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.945046] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "refresh_cache-9b539a94-6c39-46f8-b194-27047245d1f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.945458] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Instance network_info: |[{"id": "487bb20b-a111-4582-922e-c52404bb758a", "address": "fa:16:3e:85:8f:72", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487bb20b-a1", "ovs_interfaceid": "487bb20b-a111-4582-922e-c52404bb758a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.945878] env[61986]: DEBUG oslo_concurrency.lockutils [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] Acquired lock "refresh_cache-9b539a94-6c39-46f8-b194-27047245d1f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.946154] env[61986]: DEBUG nova.network.neutron [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Refreshing network info cache for port 487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.947698] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:8f:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '487bb20b-a111-4582-922e-c52404bb758a', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.957081] env[61986]: DEBUG oslo.service.loopingcall [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.957786] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 881.959034] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56ff7d63-1244-49a0-b72b-ff8e2bc59c06 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.985306] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.985306] env[61986]: value = "task-1159853" [ 881.985306] env[61986]: _type = "Task" [ 881.985306] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.997019] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159853, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.010578] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159851, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.120429] env[61986]: INFO nova.compute.manager [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Took 40.65 seconds to build instance. [ 882.153557] env[61986]: INFO nova.compute.manager [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Took 42.80 seconds to build instance. [ 882.347939] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.348641] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.351497] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.639s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.353121] env[61986]: INFO nova.compute.claims [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.499085] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159853, 'name': CreateVM_Task, 'duration_secs': 0.396287} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.507790] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 882.509046] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.509161] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.509517] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.510461] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4de62543-0d60-4485-9f1b-1083b3eabb94 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.516897] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159851, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.519294] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 882.519294] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f18cb-4bc9-96e3-27f8-bf1b583908de" [ 882.519294] env[61986]: _type = "Task" [ 882.519294] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.528442] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f18cb-4bc9-96e3-27f8-bf1b583908de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.624805] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77fadd75-c599-43c5-8456-619c7b0c44bf tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "98337bb6-9502-4d4c-af00-028659b246bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.337s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.658670] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9b31c888-27b4-4a92-ae10-e83d6cd9c959 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "4f13b629-e2a7-4668-9d77-eb638078e246" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.218s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.798641] env[61986]: DEBUG nova.network.neutron [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Updated VIF entry in instance network info cache for port 487bb20b-a111-4582-922e-c52404bb758a. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 882.799077] env[61986]: DEBUG nova.network.neutron [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Updating instance_info_cache with network_info: [{"id": "487bb20b-a111-4582-922e-c52404bb758a", "address": "fa:16:3e:85:8f:72", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487bb20b-a1", "ovs_interfaceid": "487bb20b-a111-4582-922e-c52404bb758a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.859275] env[61986]: DEBUG nova.compute.utils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.859887] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.860536] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 882.911060] env[61986]: DEBUG nova.policy [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb4b3a6fb1c7426796f38978b04940d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdab411a34424bc185f8a5981809437c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 883.010891] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159851, 'name': CloneVM_Task, 'duration_secs': 1.36823} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.011202] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created linked-clone VM from snapshot [ 883.012053] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30171908-ea5b-4939-8c9e-441e0f7181e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.025034] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Uploading image 46c39e0f-d3b3-49a0-8ab7-9754db5273df {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 883.037335] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f18cb-4bc9-96e3-27f8-bf1b583908de, 'name': SearchDatastore_Task, 'duration_secs': 0.018657} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.037645] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.038264] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.038264] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.038378] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.038476] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.038749] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0048188b-b1ee-420b-aa31-1b448ce04156 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.049195] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.049428] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 883.052975] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b232b8f-1299-4e75-8953-0f7615b75c0e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.060923] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 883.060923] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5213f529-2f30-7c25-8123-a9eb4284dfde" [ 883.060923] env[61986]: _type = "Task" [ 883.060923] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.071856] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 883.071856] env[61986]: value = "vm-252369" [ 883.071856] env[61986]: _type = "VirtualMachine" [ 883.071856] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 883.072324] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-525d76d0-cd55-4d37-8def-790f0c63c5d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.082235] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5213f529-2f30-7c25-8123-a9eb4284dfde, 'name': SearchDatastore_Task, 'duration_secs': 0.017683} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.085933] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease: (returnval){ [ 883.085933] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52daa465-fc45-2a65-5b23-ce8e28f416ae" [ 883.085933] env[61986]: _type = "HttpNfcLease" [ 883.085933] env[61986]: } obtained for exporting VM: (result){ [ 883.085933] env[61986]: value = "vm-252369" [ 883.085933] env[61986]: _type = "VirtualMachine" [ 883.085933] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 883.085933] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the lease: (returnval){ [ 883.085933] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52daa465-fc45-2a65-5b23-ce8e28f416ae" [ 883.085933] env[61986]: _type = "HttpNfcLease" [ 883.085933] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 883.085933] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8916699b-5747-4429-8b81-ec84e7c1a820 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.095672] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.095672] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52daa465-fc45-2a65-5b23-ce8e28f416ae" [ 883.095672] env[61986]: _type = "HttpNfcLease" [ 883.095672] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 883.095989] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 883.095989] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52730364-6109-4ac5-e823-3235bfa8d60c" [ 883.095989] env[61986]: _type = "Task" [ 883.095989] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.096286] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 883.096286] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52daa465-fc45-2a65-5b23-ce8e28f416ae" [ 883.096286] env[61986]: _type = "HttpNfcLease" [ 883.096286] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 883.097087] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d431cd3a-9028-4c40-8566-5e1d0a40fea2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.108121] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52730364-6109-4ac5-e823-3235bfa8d60c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.111374] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d7931a-31bc-d56a-4a12-9db0566cc1b9/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 883.111516] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d7931a-31bc-d56a-4a12-9db0566cc1b9/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 883.171877] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.175385] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.214430] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-07aa44d8-4814-41da-b7e0-4635f434c2d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.292970] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquiring lock "5170aa51-3307-42b1-b0dd-645dd4036e5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.292970] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.304063] env[61986]: DEBUG oslo_concurrency.lockutils [req-c529804e-4575-43f1-a817-0ff92b2dbb82 req-ab0be1d6-bed2-471c-bfed-b6892361629e service nova] Releasing lock "refresh_cache-9b539a94-6c39-46f8-b194-27047245d1f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.359268] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Successfully created port: 2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.374709] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.614700] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52730364-6109-4ac5-e823-3235bfa8d60c, 'name': SearchDatastore_Task, 'duration_secs': 0.029197} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.614994] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.615269] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 9b539a94-6c39-46f8-b194-27047245d1f2/9b539a94-6c39-46f8-b194-27047245d1f2.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.615553] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65bc059f-d50d-4da9-a17d-15611b63cdb6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.632040] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 883.632040] env[61986]: value = "task-1159855" [ 883.632040] env[61986]: _type = "Task" [ 883.632040] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.642113] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.647543] env[61986]: DEBUG nova.compute.manager [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Received event network-changed-494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.647891] env[61986]: DEBUG nova.compute.manager [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Refreshing instance network info cache due to event network-changed-494be155-c4b7-446d-b2ca-d49f22c7cc85. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.648208] env[61986]: DEBUG oslo_concurrency.lockutils [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] Acquiring lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.648480] env[61986]: DEBUG oslo_concurrency.lockutils [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] Acquired lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.649215] env[61986]: DEBUG nova.network.neutron [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Refreshing network info cache for port 494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.651522] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 883.651736] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252367', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'name': 'volume-853747f9-2ca2-43a2-acfb-5d955470ace1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9e57f8f9-6e9e-45fb-91d1-132490e930ae', 'attached_at': '', 'detached_at': '', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'serial': '853747f9-2ca2-43a2-acfb-5d955470ace1'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 883.653156] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a683dbac-0b7e-46b8-a3bd-0b293423db6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.676647] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec164f64-2165-4a82-a4ff-057e65e07af7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.712306] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] volume-853747f9-2ca2-43a2-acfb-5d955470ace1/volume-853747f9-2ca2-43a2-acfb-5d955470ace1.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.716404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.717493] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.718251] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a49620d6-a1b9-4f78-8a9c-3313ba00ebba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.743197] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 883.743197] env[61986]: value = "task-1159856" [ 883.743197] env[61986]: _type = "Task" [ 883.743197] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.757600] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159856, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.998088] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef901c1a-03f6-4848-9d0f-3614c4af6831 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.004221] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c4827c-47c9-4f70-b43b-fabf6638e029 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.013528] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Suspending the VM {{(pid=61986) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 884.016522] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-177bb97f-26f7-4755-b6b5-d4bc40e55122 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.020924] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd714223-3f2a-49f6-83ae-a00e989522c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.067537] env[61986]: DEBUG oslo_vmware.api [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] Waiting for the task: (returnval){ [ 884.067537] env[61986]: value = "task-1159857" [ 884.067537] env[61986]: _type = "Task" [ 884.067537] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.068547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96abf623-d3d9-42bb-a7c7-021f13e714fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.084918] env[61986]: DEBUG oslo_vmware.api [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] Task: {'id': task-1159857, 'name': SuspendVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.090506] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc88a53-3fa0-4ae6-9110-f92da28281d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.113675] env[61986]: DEBUG nova.compute.provider_tree [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.149732] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159855, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.267667] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159856, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.387674] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.414379] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.414636] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.414799] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.414992] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.415193] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.415340] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.415539] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.415701] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.415896] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.416084] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.416878] env[61986]: DEBUG nova.virt.hardware [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.417556] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1420392c-ee2a-4638-bdaa-5672944a3795 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.429964] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58536b08-4ede-4eea-8d00-80f98e27f8c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.488285] env[61986]: DEBUG nova.network.neutron [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updated VIF entry in instance network info cache for port 494be155-c4b7-446d-b2ca-d49f22c7cc85. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.488773] env[61986]: DEBUG nova.network.neutron [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updating instance_info_cache with network_info: [{"id": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "address": "fa:16:3e:85:11:df", "network": {"id": "10ba456d-8a3c-4689-a339-7b9e216a2680", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1150131841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "452cd8b9693849218d720a05cf3248c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494be155-c4", "ovs_interfaceid": "494be155-c4b7-446d-b2ca-d49f22c7cc85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.583236] env[61986]: DEBUG oslo_vmware.api [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] Task: {'id': task-1159857, 'name': SuspendVM_Task} progress is 45%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.617786] env[61986]: DEBUG nova.scheduler.client.report [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.646840] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669752} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.647479] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 9b539a94-6c39-46f8-b194-27047245d1f2/9b539a94-6c39-46f8-b194-27047245d1f2.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.647752] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.648180] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8522324c-03a2-4deb-9451-ff3ac2116e70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.657208] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 884.657208] env[61986]: value = "task-1159858" [ 884.657208] env[61986]: _type = "Task" [ 884.657208] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.757955] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159856, 'name': ReconfigVM_Task, 'duration_secs': 0.915777} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.758412] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Reconfigured VM instance instance-00000034 to attach disk [datastore1] volume-853747f9-2ca2-43a2-acfb-5d955470ace1/volume-853747f9-2ca2-43a2-acfb-5d955470ace1.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.763483] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-094a414f-7033-4e18-bc12-70342debcf9b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.782343] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 884.782343] env[61986]: value = "task-1159859" [ 884.782343] env[61986]: _type = "Task" [ 884.782343] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.792954] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.894459] env[61986]: DEBUG nova.compute.manager [req-25e378fc-f826-4603-bd70-23bcd276f927 req-fe438d56-60ad-4f14-975a-e71c9eb96963 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Received event network-vif-plugged-2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.898658] env[61986]: DEBUG oslo_concurrency.lockutils [req-25e378fc-f826-4603-bd70-23bcd276f927 req-fe438d56-60ad-4f14-975a-e71c9eb96963 service nova] Acquiring lock "d72cd38b-ef14-467b-bf53-97d9e66534c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.899134] env[61986]: DEBUG oslo_concurrency.lockutils [req-25e378fc-f826-4603-bd70-23bcd276f927 req-fe438d56-60ad-4f14-975a-e71c9eb96963 service nova] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.899491] env[61986]: DEBUG oslo_concurrency.lockutils [req-25e378fc-f826-4603-bd70-23bcd276f927 req-fe438d56-60ad-4f14-975a-e71c9eb96963 service nova] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.899833] env[61986]: DEBUG nova.compute.manager [req-25e378fc-f826-4603-bd70-23bcd276f927 req-fe438d56-60ad-4f14-975a-e71c9eb96963 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] No waiting events found dispatching network-vif-plugged-2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.900196] env[61986]: WARNING nova.compute.manager [req-25e378fc-f826-4603-bd70-23bcd276f927 req-fe438d56-60ad-4f14-975a-e71c9eb96963 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Received unexpected event network-vif-plugged-2ff255dc-13d2-4fe1-8721-ccad82b89ff1 for instance with vm_state building and task_state spawning. [ 884.992482] env[61986]: DEBUG oslo_concurrency.lockutils [req-226ba606-3931-4c73-acff-a2f1a93b09c9 req-34725438-d01e-46ba-a972-9d2d8ba77de7 service nova] Releasing lock "refresh_cache-98337bb6-9502-4d4c-af00-028659b246bf" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.005361] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Successfully updated port: 2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.084047] env[61986]: DEBUG oslo_vmware.api [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] Task: {'id': task-1159857, 'name': SuspendVM_Task, 'duration_secs': 0.923053} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.085024] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Suspended the VM {{(pid=61986) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 885.085024] env[61986]: DEBUG nova.compute.manager [None req-40fcac9f-7b94-4e4e-a9d5-67519ee5e33b tempest-ServersAdminNegativeTestJSON-2098671033 tempest-ServersAdminNegativeTestJSON-2098671033-project-admin] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.085767] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c3200c-c7b7-4436-b195-e5b42ea9236f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.122928] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.771s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.123518] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.127675] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.626s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.129836] env[61986]: INFO nova.compute.claims [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.169938] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090594} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.170900] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.171270] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fa4158-afa3-443a-b616-3b8ce26da41f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.196031] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 9b539a94-6c39-46f8-b194-27047245d1f2/9b539a94-6c39-46f8-b194-27047245d1f2.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.196674] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd209993-bc9c-4369-8a3a-a0105ce35350 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.218637] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 885.218637] env[61986]: value = "task-1159860" [ 885.218637] env[61986]: _type = "Task" [ 885.218637] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.228278] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.293291] env[61986]: DEBUG oslo_vmware.api [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1159859, 'name': ReconfigVM_Task, 'duration_secs': 0.199779} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.293873] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252367', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'name': 'volume-853747f9-2ca2-43a2-acfb-5d955470ace1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9e57f8f9-6e9e-45fb-91d1-132490e930ae', 'attached_at': '', 'detached_at': '', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'serial': '853747f9-2ca2-43a2-acfb-5d955470ace1'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 885.509133] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "refresh_cache-d72cd38b-ef14-467b-bf53-97d9e66534c8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.509133] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "refresh_cache-d72cd38b-ef14-467b-bf53-97d9e66534c8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.509133] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.635137] env[61986]: DEBUG nova.compute.utils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.639868] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.640381] env[61986]: DEBUG nova.network.neutron [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 885.692772] env[61986]: DEBUG nova.policy [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c93b2022a554633b279dba6354f23f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50561f3dfae840348c86bcce568d354f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 885.730723] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.970755] env[61986]: DEBUG nova.network.neutron [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Successfully created port: e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.064201] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.146397] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.233780] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159860, 'name': ReconfigVM_Task, 'duration_secs': 0.780427} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.234341] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 9b539a94-6c39-46f8-b194-27047245d1f2/9b539a94-6c39-46f8-b194-27047245d1f2.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.237404] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e2aca7a-5e07-4c8e-8d90-d6eb7dd0acde {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.252026] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 886.252026] env[61986]: value = "task-1159861" [ 886.252026] env[61986]: _type = "Task" [ 886.252026] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.263030] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159861, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.311609] env[61986]: DEBUG nova.network.neutron [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Updating instance_info_cache with network_info: [{"id": "2ff255dc-13d2-4fe1-8721-ccad82b89ff1", "address": "fa:16:3e:56:db:92", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff255dc-13", "ovs_interfaceid": "2ff255dc-13d2-4fe1-8721-ccad82b89ff1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.349131] env[61986]: DEBUG nova.objects.instance [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'flavor' on Instance uuid 9e57f8f9-6e9e-45fb-91d1-132490e930ae {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.659153] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548cc777-e056-4666-9261-f856d46916b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.668773] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f579abfa-20c1-4d9b-8874-d1c3a0f6d33d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.706923] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e06c30-28f6-41be-81d5-1b5b99126d32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.717605] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbc8fd4-79ba-44dd-9241-f4f20f7b06e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.733240] env[61986]: DEBUG nova.compute.provider_tree [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.762301] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159861, 'name': Rename_Task, 'duration_secs': 0.221366} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.762711] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 886.763075] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96fa3c87-ce63-436f-9c7a-8b7d265688ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.773053] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 886.773053] env[61986]: value = "task-1159862" [ 886.773053] env[61986]: _type = "Task" [ 886.773053] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.783052] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159862, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.813665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "refresh_cache-d72cd38b-ef14-467b-bf53-97d9e66534c8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.814235] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Instance network_info: |[{"id": "2ff255dc-13d2-4fe1-8721-ccad82b89ff1", "address": "fa:16:3e:56:db:92", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff255dc-13", "ovs_interfaceid": "2ff255dc-13d2-4fe1-8721-ccad82b89ff1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 886.815029] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:db:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0746f464-a938-427b-ba02-600449df5070', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ff255dc-13d2-4fe1-8721-ccad82b89ff1', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.825552] env[61986]: DEBUG oslo.service.loopingcall [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.825872] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 886.826223] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4df7a52e-fe7a-4c79-823f-718e0b2e1da8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.850557] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.850557] env[61986]: value = "task-1159863" [ 886.850557] env[61986]: _type = "Task" [ 886.850557] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.857210] env[61986]: DEBUG oslo_concurrency.lockutils [None req-36f7459e-a416-4735-ac69-3f1885921ad8 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.854s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.861668] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159863, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.987480] env[61986]: DEBUG nova.compute.manager [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Received event network-changed-2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.987690] env[61986]: DEBUG nova.compute.manager [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Refreshing instance network info cache due to event network-changed-2ff255dc-13d2-4fe1-8721-ccad82b89ff1. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.987917] env[61986]: DEBUG oslo_concurrency.lockutils [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] Acquiring lock "refresh_cache-d72cd38b-ef14-467b-bf53-97d9e66534c8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.990106] env[61986]: DEBUG oslo_concurrency.lockutils [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] Acquired lock "refresh_cache-d72cd38b-ef14-467b-bf53-97d9e66534c8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.990366] env[61986]: DEBUG nova.network.neutron [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Refreshing network info cache for port 2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 887.164430] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.192262] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.192601] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.192809] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.193224] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.193564] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.193774] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.194046] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.194477] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.194477] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.194726] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.194945] env[61986]: DEBUG nova.virt.hardware [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.196122] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57d0868-dfd5-4554-9b11-27579190e83c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.206306] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4006749-b81a-47fb-aa2c-a88a5f371292 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.236779] env[61986]: DEBUG nova.scheduler.client.report [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.284662] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159862, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.362159] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159863, 'name': CreateVM_Task, 'duration_secs': 0.432008} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.362627] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.363404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.363592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.363963] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.364284] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26a5ce78-b32e-48ce-b269-51c1f9a25c98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.370959] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 887.370959] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a464c4-17d2-75e2-30e7-c725582c24e3" [ 887.370959] env[61986]: _type = "Task" [ 887.370959] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.380607] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a464c4-17d2-75e2-30e7-c725582c24e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.595882] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "4f13b629-e2a7-4668-9d77-eb638078e246" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.596288] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "4f13b629-e2a7-4668-9d77-eb638078e246" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.596589] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "4f13b629-e2a7-4668-9d77-eb638078e246-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.596856] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "4f13b629-e2a7-4668-9d77-eb638078e246-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.597119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "4f13b629-e2a7-4668-9d77-eb638078e246-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.599654] env[61986]: INFO nova.compute.manager [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Terminating instance [ 887.601962] env[61986]: DEBUG nova.compute.manager [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.602236] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.603229] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea56b5b-f3fe-4fff-b816-78363de3defb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.613220] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 887.613545] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53a55ebd-a5cb-45be-84a8-7cfa781a2001 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.692828] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 887.693159] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 887.693394] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Deleting the datastore file [datastore2] 4f13b629-e2a7-4668-9d77-eb638078e246 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.693749] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccd67c00-3652-4eb6-a95f-6032114631ec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.702348] env[61986]: DEBUG oslo_vmware.api [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 887.702348] env[61986]: value = "task-1159865" [ 887.702348] env[61986]: _type = "Task" [ 887.702348] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.712444] env[61986]: DEBUG oslo_vmware.api [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.742077] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.742656] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.745507] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.444s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.745773] env[61986]: DEBUG nova.objects.instance [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lazy-loading 'resources' on Instance uuid 6f38006d-b1f5-4aeb-9124-1e32378c22b2 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.784277] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159862, 'name': PowerOnVM_Task, 'duration_secs': 0.593306} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.784570] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 887.784777] env[61986]: INFO nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Took 8.18 seconds to spawn the instance on the hypervisor. [ 887.784971] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.785806] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffaeac6b-a7ec-4803-93e0-f6e45972c965 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.885701] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a464c4-17d2-75e2-30e7-c725582c24e3, 'name': SearchDatastore_Task, 'duration_secs': 0.01667} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.886054] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.888031] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.888031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.888031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.888031] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.888031] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-326330f2-dabb-44a0-8c6e-d6c5a45afa24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.897374] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.897576] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 887.898412] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9389ccd0-1758-48c0-80a7-ee5cca50aa22 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.905414] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 887.905414] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5242eb79-2c50-4ebe-01f3-f9fa18588b76" [ 887.905414] env[61986]: _type = "Task" [ 887.905414] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.915942] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5242eb79-2c50-4ebe-01f3-f9fa18588b76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.928640] env[61986]: DEBUG nova.compute.manager [req-a1d77c0b-5c21-4e21-bb42-b8a9854f6ca7 req-122d6158-a44f-4362-8f6f-568541327aa7 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Received event network-vif-plugged-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.928951] env[61986]: DEBUG oslo_concurrency.lockutils [req-a1d77c0b-5c21-4e21-bb42-b8a9854f6ca7 req-122d6158-a44f-4362-8f6f-568541327aa7 service nova] Acquiring lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.929204] env[61986]: DEBUG oslo_concurrency.lockutils [req-a1d77c0b-5c21-4e21-bb42-b8a9854f6ca7 req-122d6158-a44f-4362-8f6f-568541327aa7 service nova] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.929480] env[61986]: DEBUG oslo_concurrency.lockutils [req-a1d77c0b-5c21-4e21-bb42-b8a9854f6ca7 req-122d6158-a44f-4362-8f6f-568541327aa7 service nova] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.929733] env[61986]: DEBUG nova.compute.manager [req-a1d77c0b-5c21-4e21-bb42-b8a9854f6ca7 req-122d6158-a44f-4362-8f6f-568541327aa7 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] No waiting events found dispatching network-vif-plugged-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.929927] env[61986]: WARNING nova.compute.manager [req-a1d77c0b-5c21-4e21-bb42-b8a9854f6ca7 req-122d6158-a44f-4362-8f6f-568541327aa7 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Received unexpected event network-vif-plugged-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 for instance with vm_state building and task_state spawning. [ 888.044050] env[61986]: DEBUG nova.network.neutron [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Updated VIF entry in instance network info cache for port 2ff255dc-13d2-4fe1-8721-ccad82b89ff1. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.044450] env[61986]: DEBUG nova.network.neutron [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Updating instance_info_cache with network_info: [{"id": "2ff255dc-13d2-4fe1-8721-ccad82b89ff1", "address": "fa:16:3e:56:db:92", "network": {"id": "5337888c-8a9b-42a3-bfc9-a1ec54568f10", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1253669034-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdab411a34424bc185f8a5981809437c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0746f464-a938-427b-ba02-600449df5070", "external-id": "nsx-vlan-transportzone-881", "segmentation_id": 881, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff255dc-13", "ovs_interfaceid": "2ff255dc-13d2-4fe1-8721-ccad82b89ff1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.213545] env[61986]: DEBUG oslo_vmware.api [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159865, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321162} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.214051] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.214187] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 888.214323] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 888.214631] env[61986]: INFO nova.compute.manager [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Took 0.61 seconds to destroy the instance on the hypervisor. [ 888.214873] env[61986]: DEBUG oslo.service.loopingcall [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.215116] env[61986]: DEBUG nova.compute.manager [-] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.215218] env[61986]: DEBUG nova.network.neutron [-] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 888.249247] env[61986]: DEBUG nova.compute.utils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.253622] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.253813] env[61986]: DEBUG nova.network.neutron [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 888.310123] env[61986]: INFO nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Took 39.40 seconds to build instance. [ 888.355098] env[61986]: DEBUG nova.policy [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6881bdd48c4f0ea6bcb7d2d3746d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e9d0c35a4de4f5f9829a7f3c88fde92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 888.417687] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5242eb79-2c50-4ebe-01f3-f9fa18588b76, 'name': SearchDatastore_Task, 'duration_secs': 0.016621} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.418747] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b01be66-05a6-4877-a74e-ed991bcf6fcf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.425516] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 888.425516] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5271f848-1472-2b93-c9c2-dba7d053ec54" [ 888.425516] env[61986]: _type = "Task" [ 888.425516] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.436236] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5271f848-1472-2b93-c9c2-dba7d053ec54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.547824] env[61986]: DEBUG oslo_concurrency.lockutils [req-9a5665d4-b143-46fe-a9d2-9a3cda55f905 req-f5b50a81-853e-486c-90dd-53af0b0a2f39 service nova] Releasing lock "refresh_cache-d72cd38b-ef14-467b-bf53-97d9e66534c8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.553768] env[61986]: DEBUG nova.network.neutron [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Successfully updated port: e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.753235] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c66508-eb7a-4202-bc01-62c8c54495a7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.756536] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 888.767329] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cf8d3c-2118-42e8-808e-90c787a015e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.821034] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "9b539a94-6c39-46f8-b194-27047245d1f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.469s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.827150] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034872b2-595f-48b6-b393-454f90344853 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.840543] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b89cd1-6d35-4bb2-b871-a43aa2613c07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.857128] env[61986]: DEBUG nova.compute.provider_tree [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.930668] env[61986]: DEBUG nova.network.neutron [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Successfully created port: ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.940175] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5271f848-1472-2b93-c9c2-dba7d053ec54, 'name': SearchDatastore_Task, 'duration_secs': 0.013908} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.940175] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.940490] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] d72cd38b-ef14-467b-bf53-97d9e66534c8/d72cd38b-ef14-467b-bf53-97d9e66534c8.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.940782] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6596e4ff-a53d-4401-b9d2-78c6dbda98e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.949454] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 888.949454] env[61986]: value = "task-1159866" [ 888.949454] env[61986]: _type = "Task" [ 888.949454] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.958920] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.027126] env[61986]: DEBUG nova.network.neutron [-] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.056496] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.056496] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.056496] env[61986]: DEBUG nova.network.neutron [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.333099] env[61986]: DEBUG nova.compute.manager [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Received event network-vif-deleted-7d08d53f-b8f7-4397-acce-ff7cdeeb15c0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 889.333099] env[61986]: DEBUG nova.compute.manager [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Received event network-changed-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 889.333099] env[61986]: DEBUG nova.compute.manager [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Refreshing instance network info cache due to event network-changed-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 889.333099] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] Acquiring lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.333099] env[61986]: DEBUG nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.361043] env[61986]: DEBUG nova.scheduler.client.report [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.462847] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159866, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.532353] env[61986]: INFO nova.compute.manager [-] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Took 1.32 seconds to deallocate network for instance. [ 889.594182] env[61986]: DEBUG nova.network.neutron [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.610036] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.610330] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.749062] env[61986]: DEBUG nova.network.neutron [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [{"id": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "address": "fa:16:3e:98:be:f5", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape62bf9a8-ba", "ovs_interfaceid": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.768535] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.797240] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='fc3f0bbb2be55e412f100ce3c160b10d',container_format='bare',created_at=2024-10-07T06:42:54Z,direct_url=,disk_format='vmdk',id=486535fc-ed79-4f4f-a6fb-1cb442ccd708,min_disk=1,min_ram=0,name='tempest-test-snap-1045047845',owner='4e9d0c35a4de4f5f9829a7f3c88fde92',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-07T06:43:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.798027] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.798027] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.798540] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.798722] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.798892] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.799186] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.799368] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.799577] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.799755] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.799934] env[61986]: DEBUG nova.virt.hardware [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.800857] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870f0b57-f400-44d4-be58-a8941cdeded6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.813050] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9571f3f-19c5-4c74-a7f4-4939a28aaac2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.859645] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.867782] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.122s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.870799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.577s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.872600] env[61986]: INFO nova.compute.claims [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.890329] env[61986]: INFO nova.scheduler.client.report [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Deleted allocations for instance 6f38006d-b1f5-4aeb-9124-1e32378c22b2 [ 889.961207] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678642} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.961533] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] d72cd38b-ef14-467b-bf53-97d9e66534c8/d72cd38b-ef14-467b-bf53-97d9e66534c8.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 889.961779] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.962067] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54baca49-9c0d-41bf-b56d-dd7ee1e1bfe6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.970181] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 889.970181] env[61986]: value = "task-1159867" [ 889.970181] env[61986]: _type = "Task" [ 889.970181] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.980199] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.040631] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.250459] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.250822] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance network_info: |[{"id": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "address": "fa:16:3e:98:be:f5", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape62bf9a8-ba", "ovs_interfaceid": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.251187] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] Acquired lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.251372] env[61986]: DEBUG nova.network.neutron [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Refreshing network info cache for port e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.252629] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:be:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496ac502-bfc4-4324-8332-cac473eb7cc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e62bf9a8-ba15-4ba9-8409-812d0e3c98a0', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.260315] env[61986]: DEBUG oslo.service.loopingcall [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.263273] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 890.263752] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd84ecce-16ac-40f5-bed0-25436682c4ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.285542] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.285542] env[61986]: value = "task-1159868" [ 890.285542] env[61986]: _type = "Task" [ 890.285542] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.296757] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159868, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.398508] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0852a1c6-9e80-44c3-9606-540168019da0 tempest-VolumesAdminNegativeTest-68264507 tempest-VolumesAdminNegativeTest-68264507-project-member] Lock "6f38006d-b1f5-4aeb-9124-1e32378c22b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.090s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.481723] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076894} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.482025] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.482850] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b862de72-d184-41ff-ba60-fe6f7b9cb655 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.506645] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] d72cd38b-ef14-467b-bf53-97d9e66534c8/d72cd38b-ef14-467b-bf53-97d9e66534c8.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.508283] env[61986]: DEBUG nova.network.neutron [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updated VIF entry in instance network info cache for port e62bf9a8-ba15-4ba9-8409-812d0e3c98a0. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 890.508551] env[61986]: DEBUG nova.network.neutron [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [{"id": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "address": "fa:16:3e:98:be:f5", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape62bf9a8-ba", "ovs_interfaceid": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.509781] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-840c63e2-3536-447d-a88c-bd162016dc75 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.534859] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 890.534859] env[61986]: value = "task-1159869" [ 890.534859] env[61986]: _type = "Task" [ 890.534859] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.549974] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159869, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.735865] env[61986]: DEBUG nova.network.neutron [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Successfully updated port: ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.748143] env[61986]: DEBUG nova.compute.manager [req-495e7ae4-2a88-4156-9fee-7522cefd4cca req-2ffca32b-90f5-476b-bd5c-00aec1128160 service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Received event network-vif-plugged-ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.748599] env[61986]: DEBUG oslo_concurrency.lockutils [req-495e7ae4-2a88-4156-9fee-7522cefd4cca req-2ffca32b-90f5-476b-bd5c-00aec1128160 service nova] Acquiring lock "6f9349b7-886a-4077-8f6d-a9800ab353fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.748938] env[61986]: DEBUG oslo_concurrency.lockutils [req-495e7ae4-2a88-4156-9fee-7522cefd4cca req-2ffca32b-90f5-476b-bd5c-00aec1128160 service nova] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.750085] env[61986]: DEBUG oslo_concurrency.lockutils [req-495e7ae4-2a88-4156-9fee-7522cefd4cca req-2ffca32b-90f5-476b-bd5c-00aec1128160 service nova] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.750500] env[61986]: DEBUG nova.compute.manager [req-495e7ae4-2a88-4156-9fee-7522cefd4cca req-2ffca32b-90f5-476b-bd5c-00aec1128160 service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] No waiting events found dispatching network-vif-plugged-ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.750884] env[61986]: WARNING nova.compute.manager [req-495e7ae4-2a88-4156-9fee-7522cefd4cca req-2ffca32b-90f5-476b-bd5c-00aec1128160 service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Received unexpected event network-vif-plugged-ddc2cc86-1e6d-4436-b6f1-5c967d85c128 for instance with vm_state building and task_state spawning. [ 890.794060] env[61986]: DEBUG nova.compute.manager [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Received event network-changed-ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.794599] env[61986]: DEBUG nova.compute.manager [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Refreshing instance network info cache due to event network-changed-ddc2cc86-1e6d-4436-b6f1-5c967d85c128. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 890.795016] env[61986]: DEBUG oslo_concurrency.lockutils [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] Acquiring lock "refresh_cache-6f9349b7-886a-4077-8f6d-a9800ab353fe" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.795306] env[61986]: DEBUG oslo_concurrency.lockutils [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] Acquired lock "refresh_cache-6f9349b7-886a-4077-8f6d-a9800ab353fe" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.795615] env[61986]: DEBUG nova.network.neutron [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Refreshing network info cache for port ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.802139] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159868, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.027287] env[61986]: DEBUG oslo_concurrency.lockutils [req-0a874801-21ed-403d-8282-6c5f80a96811 req-975665a9-8d3d-4429-ae2c-213727bba28d service nova] Releasing lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.045455] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159869, 'name': ReconfigVM_Task, 'duration_secs': 0.333917} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.045751] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Reconfigured VM instance instance-00000042 to attach disk [datastore2] d72cd38b-ef14-467b-bf53-97d9e66534c8/d72cd38b-ef14-467b-bf53-97d9e66534c8.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.046418] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b9a31da-dd4e-479e-8e05-25ecec2d487f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.053752] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 891.053752] env[61986]: value = "task-1159870" [ 891.053752] env[61986]: _type = "Task" [ 891.053752] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.064901] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159870, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.238819] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-6f9349b7-886a-4077-8f6d-a9800ab353fe" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.299182] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159868, 'name': CreateVM_Task, 'duration_secs': 0.586008} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.299591] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 891.302844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.303047] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.303404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.306191] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c2f5b3a-0f8f-46de-a3fa-4e28dc18f9f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.311901] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 891.311901] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52025aac-9e3c-ea7d-5639-5015eec06baf" [ 891.311901] env[61986]: _type = "Task" [ 891.311901] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.325222] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52025aac-9e3c-ea7d-5639-5015eec06baf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.379843] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed76cdbc-340b-460b-a8cd-30a66e553756 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.389167] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cd522d-3fa3-4f98-9148-b6fae6898bd9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.396468] env[61986]: DEBUG nova.network.neutron [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 891.430714] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61e1f0e-470e-418a-ba73-5a35201dfb7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.440290] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5e7070-fbcc-4e2a-bc55-3f81f95eb977 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.457887] env[61986]: DEBUG nova.compute.provider_tree [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.565207] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159870, 'name': Rename_Task, 'duration_secs': 0.26257} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.565505] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.565748] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc4e37f3-f825-4254-8176-55ef4d177cf2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.574453] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 891.574453] env[61986]: value = "task-1159871" [ 891.574453] env[61986]: _type = "Task" [ 891.574453] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.584742] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.729075] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d7931a-31bc-d56a-4a12-9db0566cc1b9/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 891.729890] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d24edf-c933-42af-aebd-b007fca30770 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.737107] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d7931a-31bc-d56a-4a12-9db0566cc1b9/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 891.737292] env[61986]: ERROR oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d7931a-31bc-d56a-4a12-9db0566cc1b9/disk-0.vmdk due to incomplete transfer. [ 891.737534] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3ee217ed-b360-4a16-9332-5979ee0d9b63 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.745999] env[61986]: DEBUG oslo_vmware.rw_handles [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52d7931a-31bc-d56a-4a12-9db0566cc1b9/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 891.746221] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Uploaded image 46c39e0f-d3b3-49a0-8ab7-9754db5273df to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 891.748618] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 891.748883] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-68a30159-0aa5-47fd-9845-7234f91e733d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.756278] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 891.756278] env[61986]: value = "task-1159872" [ 891.756278] env[61986]: _type = "Task" [ 891.756278] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.766062] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159872, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.824992] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52025aac-9e3c-ea7d-5639-5015eec06baf, 'name': SearchDatastore_Task, 'duration_secs': 0.018829} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.825292] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.825550] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.825799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.825947] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.826158] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.826468] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a04cef39-c954-4bb9-ace5-541204b61725 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.836177] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.836401] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.837471] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ed3eed2-e692-44ba-b0cd-71d3ece27d80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.843770] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 891.843770] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520d6b12-1e35-d07f-a43e-6c89b8d7b72f" [ 891.843770] env[61986]: _type = "Task" [ 891.843770] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.852119] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520d6b12-1e35-d07f-a43e-6c89b8d7b72f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.853647] env[61986]: DEBUG nova.network.neutron [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.962035] env[61986]: DEBUG nova.scheduler.client.report [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.088513] env[61986]: DEBUG oslo_vmware.api [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159871, 'name': PowerOnVM_Task, 'duration_secs': 0.503482} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.088813] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.089035] env[61986]: INFO nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Took 7.70 seconds to spawn the instance on the hypervisor. [ 892.089225] env[61986]: DEBUG nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.090045] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da03c41a-cc72-451b-927b-a2c59a7a704b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.266696] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159872, 'name': Destroy_Task, 'duration_secs': 0.32129} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.266696] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroyed the VM [ 892.267061] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 892.267342] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bde10572-a93f-41e1-9acc-9d4f763b61e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.275168] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 892.275168] env[61986]: value = "task-1159873" [ 892.275168] env[61986]: _type = "Task" [ 892.275168] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.284438] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159873, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.360382] env[61986]: DEBUG oslo_concurrency.lockutils [req-01b39cd3-c2f3-436c-b0c7-ad77bfeb1771 req-5d15a16a-f507-4b1e-a1db-0806518ffdcc service nova] Releasing lock "refresh_cache-6f9349b7-886a-4077-8f6d-a9800ab353fe" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.361110] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520d6b12-1e35-d07f-a43e-6c89b8d7b72f, 'name': SearchDatastore_Task, 'duration_secs': 0.009365} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.361171] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-6f9349b7-886a-4077-8f6d-a9800ab353fe" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.361375] env[61986]: DEBUG nova.network.neutron [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.367324] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-620aed66-fdf1-481e-a8e0-752f3f4d61d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.379959] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 892.379959] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228afd9-dd0a-19c2-f8c6-eecd92e784f8" [ 892.379959] env[61986]: _type = "Task" [ 892.379959] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.392849] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228afd9-dd0a-19c2-f8c6-eecd92e784f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.468024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.468024] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.471301] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.703s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.472521] env[61986]: INFO nova.compute.claims [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.610774] env[61986]: INFO nova.compute.manager [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Took 33.74 seconds to build instance. [ 892.787524] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159873, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.894361] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228afd9-dd0a-19c2-f8c6-eecd92e784f8, 'name': SearchDatastore_Task, 'duration_secs': 0.015805} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.894796] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.895194] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] a101dc3c-ca6b-4a72-a9b4-051b077a10fd/a101dc3c-ca6b-4a72-a9b4-051b077a10fd.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.895721] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10143869-f736-40c1-9d98-c0fd84c29777 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.906019] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 892.906019] env[61986]: value = "task-1159874" [ 892.906019] env[61986]: _type = "Task" [ 892.906019] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.912986] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159874, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.913943] env[61986]: DEBUG nova.network.neutron [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 892.979022] env[61986]: DEBUG nova.compute.utils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.982588] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.983193] env[61986]: DEBUG nova.network.neutron [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 893.034025] env[61986]: DEBUG nova.policy [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec2095d7611945759cb02cb900a1850d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff14829c4a4741c4939c932c10074d12', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 893.116812] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1de0692a-83a0-4b24-8c59-b68a6536b5f9 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.735s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.288814] env[61986]: DEBUG oslo_vmware.api [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159873, 'name': RemoveSnapshot_Task, 'duration_secs': 0.58598} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.289116] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 893.289350] env[61986]: INFO nova.compute.manager [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 13.90 seconds to snapshot the instance on the hypervisor. [ 893.305739] env[61986]: DEBUG nova.network.neutron [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Updating instance_info_cache with network_info: [{"id": "ddc2cc86-1e6d-4436-b6f1-5c967d85c128", "address": "fa:16:3e:9e:8c:06", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddc2cc86-1e", "ovs_interfaceid": "ddc2cc86-1e6d-4436-b6f1-5c967d85c128", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.325123] env[61986]: DEBUG nova.network.neutron [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Successfully created port: ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.416090] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159874, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.487828] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.618784] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.817549] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-6f9349b7-886a-4077-8f6d-a9800ab353fe" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.817922] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Instance network_info: |[{"id": "ddc2cc86-1e6d-4436-b6f1-5c967d85c128", "address": "fa:16:3e:9e:8c:06", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddc2cc86-1e", "ovs_interfaceid": "ddc2cc86-1e6d-4436-b6f1-5c967d85c128", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.818940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:8c:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddc2cc86-1e6d-4436-b6f1-5c967d85c128', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.827517] env[61986]: DEBUG oslo.service.loopingcall [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.827697] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 893.828044] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eda3c65b-a954-4d4e-9920-df2011428afb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.854081] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.854081] env[61986]: value = "task-1159875" [ 893.854081] env[61986]: _type = "Task" [ 893.854081] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.865153] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159875, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.877112] env[61986]: DEBUG nova.compute.manager [None req-7add78e2-dbae-4ec9-9888-19f7a78b7637 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Found 1 images (rotation: 2) {{(pid=61986) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 893.915577] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159874, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7759} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.915848] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] a101dc3c-ca6b-4a72-a9b4-051b077a10fd/a101dc3c-ca6b-4a72-a9b4-051b077a10fd.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.916083] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.916346] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01a29cf2-50ef-41bd-9f7e-603b2aae07c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.926082] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 893.926082] env[61986]: value = "task-1159876" [ 893.926082] env[61986]: _type = "Task" [ 893.926082] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.937156] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.023222] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a37915-3f68-4053-9042-2c675eef58d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.036096] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6493491c-8bdf-43ac-82d0-991ebc86c769 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.072449] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706af8c7-4687-4690-b09f-2e517249a1da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.081141] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f787e5-eda7-4cd5-837a-dec8a995604e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.099293] env[61986]: DEBUG nova.compute.provider_tree [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.146921] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.274208] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "9b539a94-6c39-46f8-b194-27047245d1f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.274489] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "9b539a94-6c39-46f8-b194-27047245d1f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.275252] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "9b539a94-6c39-46f8-b194-27047245d1f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.275459] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "9b539a94-6c39-46f8-b194-27047245d1f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.275727] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "9b539a94-6c39-46f8-b194-27047245d1f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.277760] env[61986]: INFO nova.compute.manager [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Terminating instance [ 894.279660] env[61986]: DEBUG nova.compute.manager [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.279858] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.280698] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d72661-658a-4973-aef7-014ab0de6e92 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.289320] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.289621] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a523d373-b848-4bb4-bd54-a2c15508777f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.296913] env[61986]: DEBUG oslo_vmware.api [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 894.296913] env[61986]: value = "task-1159877" [ 894.296913] env[61986]: _type = "Task" [ 894.296913] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.306453] env[61986]: DEBUG oslo_vmware.api [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.367240] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159875, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.436729] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083837} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.437052] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.437925] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3030ec9e-1dd3-4eca-9ca9-85eee20cb737 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.461451] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] a101dc3c-ca6b-4a72-a9b4-051b077a10fd/a101dc3c-ca6b-4a72-a9b4-051b077a10fd.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.461705] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-464f3c69-c164-4158-b8aa-356c0a9c4e14 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.482160] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 894.482160] env[61986]: value = "task-1159878" [ 894.482160] env[61986]: _type = "Task" [ 894.482160] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.495216] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.504700] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.527328] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "d72cd38b-ef14-467b-bf53-97d9e66534c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.527629] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.527839] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "d72cd38b-ef14-467b-bf53-97d9e66534c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.527999] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.528188] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.532422] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.532738] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.532915] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.533455] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.533455] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.533594] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.533770] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.533893] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.534051] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.534224] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.534412] env[61986]: DEBUG nova.virt.hardware [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.535468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2055e656-8195-4200-8e44-746a01e43555 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.539305] env[61986]: INFO nova.compute.manager [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Terminating instance [ 894.541808] env[61986]: DEBUG nova.compute.manager [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.542027] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.542857] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e944ed7-70e4-498f-a425-83eb0a81293d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.549525] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a504ab89-a8b2-4b29-8e96-1a92dbb05132 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.555876] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.556553] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbb05ad1-5cf5-4fa8-8c34-b4b8b5a0167d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.567905] env[61986]: DEBUG oslo_vmware.api [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 894.567905] env[61986]: value = "task-1159879" [ 894.567905] env[61986]: _type = "Task" [ 894.567905] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.576806] env[61986]: DEBUG oslo_vmware.api [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.601285] env[61986]: DEBUG nova.scheduler.client.report [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.810241] env[61986]: DEBUG oslo_vmware.api [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159877, 'name': PowerOffVM_Task, 'duration_secs': 0.249526} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.810241] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 894.810241] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 894.810379] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5386f23-b24d-497b-b95c-7ef3f58faf5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.865846] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159875, 'name': CreateVM_Task, 'duration_secs': 0.677796} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.866045] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 894.866768] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.866943] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.867360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.867640] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6a5d058-88c6-4f12-8995-e76204562a31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.873373] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 894.873373] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f092a8-3000-7df2-eab6-128f858de967" [ 894.873373] env[61986]: _type = "Task" [ 894.873373] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.878888] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 894.880967] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 894.880967] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleting the datastore file [datastore1] 9b539a94-6c39-46f8-b194-27047245d1f2 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.880967] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97426830-a0f7-4d7d-bf33-addb8ec863ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.884871] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f092a8-3000-7df2-eab6-128f858de967, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.890491] env[61986]: DEBUG oslo_vmware.api [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 894.890491] env[61986]: value = "task-1159881" [ 894.890491] env[61986]: _type = "Task" [ 894.890491] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.898486] env[61986]: DEBUG oslo_vmware.api [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.963802] env[61986]: DEBUG nova.network.neutron [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Successfully updated port: ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.997391] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159878, 'name': ReconfigVM_Task, 'duration_secs': 0.306131} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.997587] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Reconfigured VM instance instance-00000043 to attach disk [datastore2] a101dc3c-ca6b-4a72-a9b4-051b077a10fd/a101dc3c-ca6b-4a72-a9b4-051b077a10fd.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.998183] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96844227-f3d3-4923-8549-25e61920d87d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.008913] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 895.008913] env[61986]: value = "task-1159882" [ 895.008913] env[61986]: _type = "Task" [ 895.008913] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.021341] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159882, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.077970] env[61986]: DEBUG oslo_vmware.api [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159879, 'name': PowerOffVM_Task, 'duration_secs': 0.213725} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.078371] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.078583] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 895.078845] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54f408a0-d5cb-4da3-9474-a37944b923ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.108608] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.109243] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.111997] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.118s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.113443] env[61986]: INFO nova.compute.claims [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.117709] env[61986]: DEBUG nova.compute.manager [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.118711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b8abd2-ea02-4525-a998-e3d80e9fe0bd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.139629] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 895.140277] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 895.140277] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleting the datastore file [datastore2] d72cd38b-ef14-467b-bf53-97d9e66534c8 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.140439] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eca940ac-2e5a-4f74-a9d8-7ee80fecdf89 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.147917] env[61986]: DEBUG oslo_vmware.api [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for the task: (returnval){ [ 895.147917] env[61986]: value = "task-1159884" [ 895.147917] env[61986]: _type = "Task" [ 895.147917] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.160480] env[61986]: DEBUG oslo_vmware.api [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.199510] env[61986]: DEBUG nova.compute.manager [req-0898ec98-2f7c-4e12-936f-d80a740dfef6 req-32489248-ddc7-4623-88fe-b754e32421d8 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Received event network-vif-plugged-ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.199733] env[61986]: DEBUG oslo_concurrency.lockutils [req-0898ec98-2f7c-4e12-936f-d80a740dfef6 req-32489248-ddc7-4623-88fe-b754e32421d8 service nova] Acquiring lock "b833adc9-2967-4b0b-81c2-0b8deac20f69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.199942] env[61986]: DEBUG oslo_concurrency.lockutils [req-0898ec98-2f7c-4e12-936f-d80a740dfef6 req-32489248-ddc7-4623-88fe-b754e32421d8 service nova] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.200127] env[61986]: DEBUG oslo_concurrency.lockutils [req-0898ec98-2f7c-4e12-936f-d80a740dfef6 req-32489248-ddc7-4623-88fe-b754e32421d8 service nova] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.200290] env[61986]: DEBUG nova.compute.manager [req-0898ec98-2f7c-4e12-936f-d80a740dfef6 req-32489248-ddc7-4623-88fe-b754e32421d8 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] No waiting events found dispatching network-vif-plugged-ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.200456] env[61986]: WARNING nova.compute.manager [req-0898ec98-2f7c-4e12-936f-d80a740dfef6 req-32489248-ddc7-4623-88fe-b754e32421d8 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Received unexpected event network-vif-plugged-ff9551fd-6a35-45fb-b841-fb3e1c73861c for instance with vm_state building and task_state spawning. [ 895.384999] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.385348] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Processing image 486535fc-ed79-4f4f-a6fb-1cb442ccd708 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.385597] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.385754] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.385936] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.386217] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-779da1d7-02d4-4cd5-8ebc-77b66d112ead {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.397799] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.398027] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 895.399122] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cae51b3-be6b-4ba4-a9a4-165dd2a0c31f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.404475] env[61986]: DEBUG oslo_vmware.api [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151982} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.405116] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.405325] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.405957] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.405957] env[61986]: INFO nova.compute.manager [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 895.406087] env[61986]: DEBUG oslo.service.loopingcall [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.406325] env[61986]: DEBUG nova.compute.manager [-] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.406462] env[61986]: DEBUG nova.network.neutron [-] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 895.409509] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 895.409509] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5219f474-e4db-dfe5-dc11-522e92bcb703" [ 895.409509] env[61986]: _type = "Task" [ 895.409509] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.421071] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5219f474-e4db-dfe5-dc11-522e92bcb703, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.467089] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "refresh_cache-b833adc9-2967-4b0b-81c2-0b8deac20f69" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.467261] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquired lock "refresh_cache-b833adc9-2967-4b0b-81c2-0b8deac20f69" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.467416] env[61986]: DEBUG nova.network.neutron [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 895.522362] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159882, 'name': Rename_Task, 'duration_secs': 0.15289} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.522648] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.523464] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5ba97c0-b6b0-4987-87cb-298e53c37909 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.531801] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 895.531801] env[61986]: value = "task-1159885" [ 895.531801] env[61986]: _type = "Task" [ 895.531801] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.550659] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159885, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.619678] env[61986]: DEBUG nova.compute.utils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.625266] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 895.626319] env[61986]: DEBUG nova.network.neutron [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 895.631665] env[61986]: INFO nova.compute.manager [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] instance snapshotting [ 895.632596] env[61986]: DEBUG nova.objects.instance [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'flavor' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.666018] env[61986]: DEBUG oslo_vmware.api [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Task: {'id': task-1159884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196061} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.666018] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.666018] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.666018] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.666018] env[61986]: INFO nova.compute.manager [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 895.666018] env[61986]: DEBUG oslo.service.loopingcall [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.666018] env[61986]: DEBUG nova.compute.manager [-] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.666018] env[61986]: DEBUG nova.network.neutron [-] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 895.694185] env[61986]: DEBUG nova.policy [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ea19b4c355d4b2797683d78e82eadea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0110dfcd30de444badf6fe685042ba3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 895.925381] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 895.925763] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Fetch image to [datastore2] OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f/OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 895.926085] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Downloading stream optimized image 486535fc-ed79-4f4f-a6fb-1cb442ccd708 to [datastore2] OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f/OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f.vmdk on the data store datastore2 as vApp {{(pid=61986) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 895.926329] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Downloading image file data 486535fc-ed79-4f4f-a6fb-1cb442ccd708 to the ESX as VM named 'OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f' {{(pid=61986) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 896.031406] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 896.031406] env[61986]: value = "resgroup-9" [ 896.031406] env[61986]: _type = "ResourcePool" [ 896.031406] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 896.031769] env[61986]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f493ff9c-c833-4cf5-a949-4535e5542929 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.052560] env[61986]: DEBUG nova.network.neutron [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Successfully created port: e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.054795] env[61986]: DEBUG nova.network.neutron [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 896.066081] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159885, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.067430] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lease: (returnval){ [ 896.067430] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b622bf-6d96-bd35-5557-3f186db3ad6a" [ 896.067430] env[61986]: _type = "HttpNfcLease" [ 896.067430] env[61986]: } obtained for vApp import into resource pool (val){ [ 896.067430] env[61986]: value = "resgroup-9" [ 896.067430] env[61986]: _type = "ResourcePool" [ 896.067430] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 896.067725] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the lease: (returnval){ [ 896.067725] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b622bf-6d96-bd35-5557-3f186db3ad6a" [ 896.067725] env[61986]: _type = "HttpNfcLease" [ 896.067725] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 896.074380] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 896.074380] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b622bf-6d96-bd35-5557-3f186db3ad6a" [ 896.074380] env[61986]: _type = "HttpNfcLease" [ 896.074380] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 896.125962] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.141292] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50cf43b1-9b5e-49e5-94f9-cf865adcbd28 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.168109] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c818e4-41a9-4731-8076-36e16cc41f92 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.298987] env[61986]: DEBUG nova.network.neutron [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Updating instance_info_cache with network_info: [{"id": "ff9551fd-6a35-45fb-b841-fb3e1c73861c", "address": "fa:16:3e:ca:6f:b9", "network": {"id": "01c3d4c7-b8a6-4fc1-9f96-a1f1357bf9b1", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-979106593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff14829c4a4741c4939c932c10074d12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff9551fd-6a", "ovs_interfaceid": "ff9551fd-6a35-45fb-b841-fb3e1c73861c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.383711] env[61986]: DEBUG nova.network.neutron [-] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.545459] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159885, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.580806] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 896.580806] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b622bf-6d96-bd35-5557-3f186db3ad6a" [ 896.580806] env[61986]: _type = "HttpNfcLease" [ 896.580806] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 896.621702] env[61986]: DEBUG nova.network.neutron [-] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.671699] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debe96f5-7edd-4ee7-b189-92168253faa2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.680497] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfd6417-7e30-4bfc-8f73-76b73e3778d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.687993] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 896.714834] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6ad1f7d4-7d30-446c-8c22-ab11d9f2ff38 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.718508] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0010c08c-f529-43fa-8923-1955de6b8c3c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.727250] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442c53dd-60f6-47de-b5b1-ad48d1badbca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.734721] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 896.734721] env[61986]: value = "task-1159887" [ 896.734721] env[61986]: _type = "Task" [ 896.734721] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.746716] env[61986]: DEBUG nova.compute.provider_tree [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.754902] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159887, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.801277] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Releasing lock "refresh_cache-b833adc9-2967-4b0b-81c2-0b8deac20f69" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.801634] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Instance network_info: |[{"id": "ff9551fd-6a35-45fb-b841-fb3e1c73861c", "address": "fa:16:3e:ca:6f:b9", "network": {"id": "01c3d4c7-b8a6-4fc1-9f96-a1f1357bf9b1", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-979106593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff14829c4a4741c4939c932c10074d12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff9551fd-6a", "ovs_interfaceid": "ff9551fd-6a35-45fb-b841-fb3e1c73861c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.802116] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:6f:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '101a44fc-ffde-4e3e-ad82-363454ae458b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff9551fd-6a35-45fb-b841-fb3e1c73861c', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.810388] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Creating folder: Project (ff14829c4a4741c4939c932c10074d12). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 896.810761] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e65d242c-52ee-4913-9742-065516259b13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.823370] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Created folder: Project (ff14829c4a4741c4939c932c10074d12) in parent group-v252271. [ 896.823615] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Creating folder: Instances. Parent ref: group-v252375. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 896.823878] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a43fe978-eb59-483d-9e3d-0328f9e3c6ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.833901] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Created folder: Instances in parent group-v252375. [ 896.834733] env[61986]: DEBUG oslo.service.loopingcall [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.834733] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 896.834733] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c86059e3-1cec-41f2-9416-3ed394c6858e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.855772] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.855772] env[61986]: value = "task-1159890" [ 896.855772] env[61986]: _type = "Task" [ 896.855772] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.863647] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159890, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.886925] env[61986]: INFO nova.compute.manager [-] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Took 1.48 seconds to deallocate network for instance. [ 897.045690] env[61986]: DEBUG oslo_vmware.api [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159885, 'name': PowerOnVM_Task, 'duration_secs': 1.135815} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.046089] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 897.046367] env[61986]: INFO nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Took 9.88 seconds to spawn the instance on the hypervisor. [ 897.046593] env[61986]: DEBUG nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.047529] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998ecbd3-2f98-46ca-96b9-8afad3979c9c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.078703] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.078703] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b622bf-6d96-bd35-5557-3f186db3ad6a" [ 897.078703] env[61986]: _type = "HttpNfcLease" [ 897.078703] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 897.078703] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 897.078703] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b622bf-6d96-bd35-5557-3f186db3ad6a" [ 897.078703] env[61986]: _type = "HttpNfcLease" [ 897.078703] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 897.079213] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22f4a15-8301-4748-be6b-73f2258441ff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.089041] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5ccf8-2dbe-6f9c-559a-671c73a19792/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 897.089041] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5ccf8-2dbe-6f9c-559a-671c73a19792/disk-0.vmdk. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 897.146767] env[61986]: INFO nova.compute.manager [-] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Took 1.48 seconds to deallocate network for instance. [ 897.147892] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.159148] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4e6f670f-7c9a-4336-a9c3-2c718bdc723f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.246896] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159887, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.249989] env[61986]: DEBUG nova.scheduler.client.report [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.288924] env[61986]: DEBUG nova.compute.manager [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Received event network-changed-ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.289750] env[61986]: DEBUG nova.compute.manager [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Refreshing instance network info cache due to event network-changed-ff9551fd-6a35-45fb-b841-fb3e1c73861c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 897.289976] env[61986]: DEBUG oslo_concurrency.lockutils [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] Acquiring lock "refresh_cache-b833adc9-2967-4b0b-81c2-0b8deac20f69" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.290225] env[61986]: DEBUG oslo_concurrency.lockutils [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] Acquired lock "refresh_cache-b833adc9-2967-4b0b-81c2-0b8deac20f69" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.290442] env[61986]: DEBUG nova.network.neutron [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Refreshing network info cache for port ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 897.311845] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.312543] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.312543] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.312543] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.312840] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.312840] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.312949] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.313153] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.313342] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.313529] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.313761] env[61986]: DEBUG nova.virt.hardware [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.315017] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16828651-cf2c-4cc5-9dd4-514dadb14078 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.324844] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d932e4-ca5e-443d-a658-2a5e49deaafa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.367083] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159890, 'name': CreateVM_Task, 'duration_secs': 0.397264} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.367276] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 897.367917] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.368115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.368419] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.368691] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7797e321-de08-4073-bfdb-cee42cfc1f5e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.374848] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 897.374848] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ad3e3d-e8b2-cfc9-ef62-1aa458b9a6ea" [ 897.374848] env[61986]: _type = "Task" [ 897.374848] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.384983] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ad3e3d-e8b2-cfc9-ef62-1aa458b9a6ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.394165] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.577430] env[61986]: INFO nova.compute.manager [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Took 34.90 seconds to build instance. [ 897.663824] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.746663] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159887, 'name': CreateSnapshot_Task, 'duration_secs': 0.543981} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.747302] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 897.748954] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56014cc-3ca1-4432-9a92-d915e9cf452a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.754518] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.642s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.755093] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.762584] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.116s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.896942] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ad3e3d-e8b2-cfc9-ef62-1aa458b9a6ea, 'name': SearchDatastore_Task, 'duration_secs': 0.012457} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.899755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.900013] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.900260] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.900413] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.900594] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.900877] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81d96b27-b3ba-436b-9307-ce0c2f19d5d2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.915492] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.915492] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 897.918403] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfff0fed-0823-435a-88f6-74ca7bddf76a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.929490] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 897.929490] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52414e14-826b-26d3-21bb-845599e605f1" [ 897.929490] env[61986]: _type = "Task" [ 897.929490] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.940432] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52414e14-826b-26d3-21bb-845599e605f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.006484] env[61986]: DEBUG nova.compute.manager [req-0b243f25-80c3-4dba-bbb0-d353f934b0f8 req-5407d387-3c07-4507-99b5-85b6f440acf7 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Received event network-vif-plugged-e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.006883] env[61986]: DEBUG oslo_concurrency.lockutils [req-0b243f25-80c3-4dba-bbb0-d353f934b0f8 req-5407d387-3c07-4507-99b5-85b6f440acf7 service nova] Acquiring lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.007180] env[61986]: DEBUG oslo_concurrency.lockutils [req-0b243f25-80c3-4dba-bbb0-d353f934b0f8 req-5407d387-3c07-4507-99b5-85b6f440acf7 service nova] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.007415] env[61986]: DEBUG oslo_concurrency.lockutils [req-0b243f25-80c3-4dba-bbb0-d353f934b0f8 req-5407d387-3c07-4507-99b5-85b6f440acf7 service nova] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.007595] env[61986]: DEBUG nova.compute.manager [req-0b243f25-80c3-4dba-bbb0-d353f934b0f8 req-5407d387-3c07-4507-99b5-85b6f440acf7 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] No waiting events found dispatching network-vif-plugged-e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.007759] env[61986]: WARNING nova.compute.manager [req-0b243f25-80c3-4dba-bbb0-d353f934b0f8 req-5407d387-3c07-4507-99b5-85b6f440acf7 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Received unexpected event network-vif-plugged-e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 for instance with vm_state building and task_state spawning. [ 898.078150] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1c33123d-05d4-457d-aab4-9b269af419fe tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.136s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.115916] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 898.116038] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5ccf8-2dbe-6f9c-559a-671c73a19792/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 898.118216] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92e569a-42f8-43ce-8cd0-0de378f358c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.124910] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5ccf8-2dbe-6f9c-559a-671c73a19792/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 898.125026] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5ccf8-2dbe-6f9c-559a-671c73a19792/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 898.125267] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7b8ee15f-fcb3-41e9-9ff4-851d65508c3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.184866] env[61986]: DEBUG nova.network.neutron [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Successfully updated port: e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.206459] env[61986]: DEBUG nova.network.neutron [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Updated VIF entry in instance network info cache for port ff9551fd-6a35-45fb-b841-fb3e1c73861c. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 898.206836] env[61986]: DEBUG nova.network.neutron [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Updating instance_info_cache with network_info: [{"id": "ff9551fd-6a35-45fb-b841-fb3e1c73861c", "address": "fa:16:3e:ca:6f:b9", "network": {"id": "01c3d4c7-b8a6-4fc1-9f96-a1f1357bf9b1", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-979106593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ff14829c4a4741c4939c932c10074d12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff9551fd-6a", "ovs_interfaceid": "ff9551fd-6a35-45fb-b841-fb3e1c73861c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.279614] env[61986]: DEBUG nova.compute.utils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.287714] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 898.288640] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.288833] env[61986]: DEBUG nova.network.neutron [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 898.295625] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.297909] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b2d02175-9c03-450c-ba37-ae7ae481960c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.312629] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 898.312629] env[61986]: value = "task-1159891" [ 898.312629] env[61986]: _type = "Task" [ 898.312629] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.323625] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159891, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.397813] env[61986]: DEBUG nova.policy [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5afe2dd2a41143a4a204e889278df0e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c94c168de68c4608b772cf6fa6012dcb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 898.443064] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52414e14-826b-26d3-21bb-845599e605f1, 'name': SearchDatastore_Task, 'duration_secs': 0.014823} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.443961] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d71f4206-9677-4fc0-8397-aa324c92575e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.451844] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 898.451844] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d05dd1-59c9-9f49-0505-d5085ccd3dc5" [ 898.451844] env[61986]: _type = "Task" [ 898.451844] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.465104] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d05dd1-59c9-9f49-0505-d5085ccd3dc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.498796] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5ccf8-2dbe-6f9c-559a-671c73a19792/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 898.499290] env[61986]: INFO nova.virt.vmwareapi.images [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Downloaded image file data 486535fc-ed79-4f4f-a6fb-1cb442ccd708 [ 898.499877] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105805f5-6769-4957-ac37-e36c3f8602e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.518037] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4d9f4bb-0f96-4006-8595-f1d43dea50bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.577198] env[61986]: INFO nova.virt.vmwareapi.images [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] The imported VM was unregistered [ 898.581429] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 898.583444] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating directory with path [datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.583444] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.587854] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d54073f-7ed0-48be-8a15-0ce469b121ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.606418] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Created directory with path [datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.606652] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f/OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f.vmdk to [datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk. {{(pid=61986) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 898.606956] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-47fbcba8-a383-453e-a2c9-65f495426e34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.615374] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 898.615374] env[61986]: value = "task-1159893" [ 898.615374] env[61986]: _type = "Task" [ 898.615374] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.624064] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.687779] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.687990] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.688214] env[61986]: DEBUG nova.network.neutron [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.710236] env[61986]: DEBUG oslo_concurrency.lockutils [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] Releasing lock "refresh_cache-b833adc9-2967-4b0b-81c2-0b8deac20f69" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.710395] env[61986]: DEBUG nova.compute.manager [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Received event network-vif-deleted-487bb20b-a111-4582-922e-c52404bb758a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.710599] env[61986]: DEBUG nova.compute.manager [req-52f963dd-90ef-4fba-be54-be1629b0777e req-ce2e11a0-8546-42f5-bfe2-99dfe3c4634c service nova] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Received event network-vif-deleted-2ff255dc-13d2-4fe1-8721-ccad82b89ff1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.821152] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa12919e-3891-4da9-a280-4155137864e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.821314] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.821437] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6e0697d4-d862-497c-9bdd-dd1ef2d4272b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.821584] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 977799b4-2793-4513-9447-483146fc7ac4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 898.821714] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance eaa61f81-2f9b-4d1c-bab8-4363fb71e936 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.821895] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 86c9e45c-b3f1-4004-9acc-190b11a4a926 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.821933] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9e57f8f9-6e9e-45fb-91d1-132490e930ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822047] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822167] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ee47cab9-5dfd-48ce-ba70-cb800d735b19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822278] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 021a27ec-cee0-454d-8daf-e6a82bd9330c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822387] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a0fa2cbe-d8c3-462a-9744-aaa36a8d314e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822509] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 46b18b6c-6e40-45b7-9d3f-6177b08a52a5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 898.822645] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 076c8e91-af70-489f-89d5-f598166fafc6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822756] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 52eda388-05f4-416c-ac9d-bd1c1e31ba9c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.822893] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance b9706e51-0f74-4dbc-aab1-ea640ac78c62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.824631] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 40de5f22-0de2-466a-91ab-dcb6ec586dad is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 898.824631] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 4f13b629-e2a7-4668-9d77-eb638078e246 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 898.824631] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 98337bb6-9502-4d4c-af00-028659b246bf actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.824631] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9b539a94-6c39-46f8-b194-27047245d1f2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 898.824631] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance d72cd38b-ef14-467b-bf53-97d9e66534c8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 898.824631] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a101dc3c-ca6b-4a72-a9b4-051b077a10fd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.824631] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6f9349b7-886a-4077-8f6d-a9800ab353fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.824631] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance b833adc9-2967-4b0b-81c2-0b8deac20f69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.824631] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance f42e92ef-8a90-473c-8662-57cbb9e3f4ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.824631] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 3512cd72-6666-4810-828e-50230956c4b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 898.829441] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159891, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.915277] env[61986]: DEBUG nova.network.neutron [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Successfully created port: 73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.964581] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d05dd1-59c9-9f49-0505-d5085ccd3dc5, 'name': SearchDatastore_Task, 'duration_secs': 0.020851} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.964888] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.965170] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] b833adc9-2967-4b0b-81c2-0b8deac20f69/b833adc9-2967-4b0b-81c2-0b8deac20f69.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 898.965462] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aedbb6fd-3b4c-4162-a69e-a9cc8eb323bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.975244] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 898.975244] env[61986]: value = "task-1159894" [ 898.975244] env[61986]: _type = "Task" [ 898.975244] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.985426] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.120679] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.129393] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.264842] env[61986]: DEBUG nova.network.neutron [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.307859] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.329893] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159891, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.333430] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 59cfee64-ef79-4b8f-a703-c8812551d12f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 899.346019] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.347403] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.347403] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.347403] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.347403] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.347640] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.347731] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.347952] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.348146] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.348294] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.348494] env[61986]: DEBUG nova.virt.hardware [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.349758] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0981f3d6-aae3-4427-ab5e-5103fe1f6fbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.365303] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8a4e5a-4d5f-4c63-8102-42f366c70f55 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.491271] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.600869] env[61986]: DEBUG nova.network.neutron [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [{"id": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "address": "fa:16:3e:5e:4f:21", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7c53e1f-f6", "ovs_interfaceid": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.636398] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.825856] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159891, 'name': CloneVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.836899] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6488f68e-7b84-4462-aef5-25d02db504f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 899.989908] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.109881] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.109881] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Instance network_info: |[{"id": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "address": "fa:16:3e:5e:4f:21", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7c53e1f-f6", "ovs_interfaceid": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.110483] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:4f:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7c53e1f-f6f4-4df9-b4b0-e2818f87b045', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.127267] env[61986]: DEBUG oslo.service.loopingcall [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.127267] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 900.135038] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ac13306-1299-46ba-8219-739242bc05bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.166559] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.168030] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.168030] env[61986]: value = "task-1159895" [ 900.168030] env[61986]: _type = "Task" [ 900.168030] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.177887] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159895, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.327050] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159891, 'name': CloneVM_Task, 'duration_secs': 1.529336} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.327924] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created linked-clone VM from snapshot [ 900.328617] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb69c984-8fbc-48e9-b4d6-bccdda02f1e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.338355] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Uploading image f56ca176-82f3-4ea2-8aaa-08ad8bc1fba2 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 900.346269] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 49bcac13-6dde-4a28-9131-f9a1e1d2e386 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 900.371120] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 900.371120] env[61986]: value = "vm-252379" [ 900.371120] env[61986]: _type = "VirtualMachine" [ 900.371120] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 900.371604] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d6c02661-4eb0-42a8-bc88-dabe58ae65e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.383774] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease: (returnval){ [ 900.383774] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52329399-4333-f6ff-84de-2876b129d850" [ 900.383774] env[61986]: _type = "HttpNfcLease" [ 900.383774] env[61986]: } obtained for exporting VM: (result){ [ 900.383774] env[61986]: value = "vm-252379" [ 900.383774] env[61986]: _type = "VirtualMachine" [ 900.383774] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 900.384468] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the lease: (returnval){ [ 900.384468] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52329399-4333-f6ff-84de-2876b129d850" [ 900.384468] env[61986]: _type = "HttpNfcLease" [ 900.384468] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 900.393865] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 900.393865] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52329399-4333-f6ff-84de-2876b129d850" [ 900.393865] env[61986]: _type = "HttpNfcLease" [ 900.393865] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 900.396702] env[61986]: DEBUG nova.compute.manager [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Received event network-changed-e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.396903] env[61986]: DEBUG nova.compute.manager [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Refreshing instance network info cache due to event network-changed-e7c53e1f-f6f4-4df9-b4b0-e2818f87b045. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.397120] env[61986]: DEBUG oslo_concurrency.lockutils [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] Acquiring lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.397678] env[61986]: DEBUG oslo_concurrency.lockutils [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] Acquired lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.397678] env[61986]: DEBUG nova.network.neutron [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Refreshing network info cache for port e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.491199] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.635574] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.679378] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159895, 'name': CreateVM_Task, 'duration_secs': 0.415037} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.683023] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 900.683023] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.683023] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.683023] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.683023] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2a07f14-811a-49a3-ac58-98254fdcd4c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.687722] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 900.687722] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d34ac3-fa1c-acb1-b52d-f32c99cf469e" [ 900.687722] env[61986]: _type = "Task" [ 900.687722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.701320] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d34ac3-fa1c-acb1-b52d-f32c99cf469e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.853925] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance dea128a9-aed0-40b2-ae17-c068ea8e3452 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 900.896699] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 900.896699] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52329399-4333-f6ff-84de-2876b129d850" [ 900.896699] env[61986]: _type = "HttpNfcLease" [ 900.896699] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 900.897084] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 900.897084] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52329399-4333-f6ff-84de-2876b129d850" [ 900.897084] env[61986]: _type = "HttpNfcLease" [ 900.897084] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 900.897803] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba3112f-c6f0-4809-b56a-9da82448cb71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.908051] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c0056b-3bd5-91e9-a229-66eefcd769a1/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 900.908277] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c0056b-3bd5-91e9-a229-66eefcd769a1/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 900.990631] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159894, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.009668] env[61986]: DEBUG nova.network.neutron [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Successfully updated port: 73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.016167] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9d64f3e7-eb0b-4669-a8fb-655dc6618406 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.142153] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.203234] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d34ac3-fa1c-acb1-b52d-f32c99cf469e, 'name': SearchDatastore_Task, 'duration_secs': 0.016373} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.206619] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.207161] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.207556] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.207727] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.207926] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.208271] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8058b928-2294-42ca-84e9-fcd2558e2cab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.221790] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.222048] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 901.222833] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5873ff5e-9b52-43e9-a770-e945698a7d4d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.230196] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 901.230196] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e26790-7a09-3f4f-bece-1482961e6a5c" [ 901.230196] env[61986]: _type = "Task" [ 901.230196] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.240952] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e26790-7a09-3f4f-bece-1482961e6a5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.358170] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 901.377883] env[61986]: DEBUG nova.network.neutron [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updated VIF entry in instance network info cache for port e7c53e1f-f6f4-4df9-b4b0-e2818f87b045. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 901.378640] env[61986]: DEBUG nova.network.neutron [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [{"id": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "address": "fa:16:3e:5e:4f:21", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7c53e1f-f6", "ovs_interfaceid": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.493773] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159894, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.378849} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.493961] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] b833adc9-2967-4b0b-81c2-0b8deac20f69/b833adc9-2967-4b0b-81c2-0b8deac20f69.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 901.494301] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.494709] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cae23540-f536-4c81-8b0d-b753d31bca2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.504058] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 901.504058] env[61986]: value = "task-1159897" [ 901.504058] env[61986]: _type = "Task" [ 901.504058] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.509113] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "refresh_cache-3512cd72-6666-4810-828e-50230956c4b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.509199] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "refresh_cache-3512cd72-6666-4810-828e-50230956c4b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.509461] env[61986]: DEBUG nova.network.neutron [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.521486] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.644019] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159893, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.760686} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.644019] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f/OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f.vmdk to [datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk. [ 901.644019] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Cleaning up location [datastore2] OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 901.644019] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_33a12c7d-59fd-4d0b-80db-7af2210cbf3f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.644019] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d43677ae-5ffe-4f82-bf77-5e2243a7c1a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.652235] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 901.652235] env[61986]: value = "task-1159898" [ 901.652235] env[61986]: _type = "Task" [ 901.652235] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.663010] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.743972] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e26790-7a09-3f4f-bece-1482961e6a5c, 'name': SearchDatastore_Task, 'duration_secs': 0.033418} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.743972] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12688d7d-0a92-408d-b56d-3a1b12141879 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.750710] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 901.750710] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c68dcf-7329-e67b-9f75-c9aa6d499a3d" [ 901.750710] env[61986]: _type = "Task" [ 901.750710] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.761245] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c68dcf-7329-e67b-9f75-c9aa6d499a3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.862208] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 3d5f4513-bbc4-404e-9d3e-340bd369fc3c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 901.881145] env[61986]: DEBUG oslo_concurrency.lockutils [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] Releasing lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.883436] env[61986]: DEBUG nova.compute.manager [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.883436] env[61986]: DEBUG nova.compute.manager [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing instance network info cache due to event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 901.883436] env[61986]: DEBUG oslo_concurrency.lockutils [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] Acquiring lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.883436] env[61986]: DEBUG oslo_concurrency.lockutils [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] Acquired lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.883436] env[61986]: DEBUG nova.network.neutron [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.004111] env[61986]: DEBUG nova.compute.manager [req-21390ea5-b32d-4a76-b49b-b00052b49889 req-0a9dba0d-baee-4886-8744-d6a53137ea18 service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Received event network-vif-plugged-73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.004111] env[61986]: DEBUG oslo_concurrency.lockutils [req-21390ea5-b32d-4a76-b49b-b00052b49889 req-0a9dba0d-baee-4886-8744-d6a53137ea18 service nova] Acquiring lock "3512cd72-6666-4810-828e-50230956c4b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.004111] env[61986]: DEBUG oslo_concurrency.lockutils [req-21390ea5-b32d-4a76-b49b-b00052b49889 req-0a9dba0d-baee-4886-8744-d6a53137ea18 service nova] Lock "3512cd72-6666-4810-828e-50230956c4b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.004111] env[61986]: DEBUG oslo_concurrency.lockutils [req-21390ea5-b32d-4a76-b49b-b00052b49889 req-0a9dba0d-baee-4886-8744-d6a53137ea18 service nova] Lock "3512cd72-6666-4810-828e-50230956c4b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.004111] env[61986]: DEBUG nova.compute.manager [req-21390ea5-b32d-4a76-b49b-b00052b49889 req-0a9dba0d-baee-4886-8744-d6a53137ea18 service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] No waiting events found dispatching network-vif-plugged-73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.004111] env[61986]: WARNING nova.compute.manager [req-21390ea5-b32d-4a76-b49b-b00052b49889 req-0a9dba0d-baee-4886-8744-d6a53137ea18 service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Received unexpected event network-vif-plugged-73b7b19d-3540-46ec-b022-6bc7aaf30b1d for instance with vm_state building and task_state spawning. [ 902.020389] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082336} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.021062] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.022373] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c376006-21ee-499f-884b-4002b08a15c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.051370] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] b833adc9-2967-4b0b-81c2-0b8deac20f69/b833adc9-2967-4b0b-81c2-0b8deac20f69.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.051902] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbb58f65-270f-47dc-ad34-01e8c103906b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.075247] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 902.075247] env[61986]: value = "task-1159899" [ 902.075247] env[61986]: _type = "Task" [ 902.075247] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.086792] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159899, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.089320] env[61986]: DEBUG nova.network.neutron [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 902.163969] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.043599} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.165151] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.165453] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "[datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.165793] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk to [datastore2] 6f9349b7-886a-4077-8f6d-a9800ab353fe/6f9349b7-886a-4077-8f6d-a9800ab353fe.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.166370] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2609fcde-c64c-4832-97de-df477e3bc368 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.176398] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 902.176398] env[61986]: value = "task-1159900" [ 902.176398] env[61986]: _type = "Task" [ 902.176398] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.190215] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.263043] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c68dcf-7329-e67b-9f75-c9aa6d499a3d, 'name': SearchDatastore_Task, 'duration_secs': 0.01193} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.263514] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.263932] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee/f42e92ef-8a90-473c-8662-57cbb9e3f4ee.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 902.264290] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e425739-b8a0-499d-8e66-881a97493b17 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.272792] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 902.272792] env[61986]: value = "task-1159901" [ 902.272792] env[61986]: _type = "Task" [ 902.272792] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.282955] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.366825] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 5170aa51-3307-42b1-b0dd-645dd4036e5b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 902.425608] env[61986]: DEBUG nova.network.neutron [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Updating instance_info_cache with network_info: [{"id": "73b7b19d-3540-46ec-b022-6bc7aaf30b1d", "address": "fa:16:3e:df:30:83", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b7b19d-35", "ovs_interfaceid": "73b7b19d-3540-46ec-b022-6bc7aaf30b1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.547045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.547441] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.590187] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159899, 'name': ReconfigVM_Task, 'duration_secs': 0.34073} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.590546] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Reconfigured VM instance instance-00000045 to attach disk [datastore2] b833adc9-2967-4b0b-81c2-0b8deac20f69/b833adc9-2967-4b0b-81c2-0b8deac20f69.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.591317] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7f18860-63bb-423b-97d4-ddd9c97acb60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.603601] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 902.603601] env[61986]: value = "task-1159902" [ 902.603601] env[61986]: _type = "Task" [ 902.603601] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.619889] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159902, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.695618] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159900, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.759309] env[61986]: DEBUG nova.network.neutron [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updated VIF entry in instance network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 902.759741] env[61986]: DEBUG nova.network.neutron [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.788977] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159901, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.870728] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance f15bf19d-b86a-4b0a-ac1a-9df8e77e9382 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 902.871112] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 902.871225] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4160MB phys_disk=200GB used_disk=18GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 902.928255] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "refresh_cache-3512cd72-6666-4810-828e-50230956c4b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.928604] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Instance network_info: |[{"id": "73b7b19d-3540-46ec-b022-6bc7aaf30b1d", "address": "fa:16:3e:df:30:83", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b7b19d-35", "ovs_interfaceid": "73b7b19d-3540-46ec-b022-6bc7aaf30b1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.929118] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:30:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e350f83a-f581-4e10-ac16-0b0f7bfd3d38', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73b7b19d-3540-46ec-b022-6bc7aaf30b1d', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.938347] env[61986]: DEBUG oslo.service.loopingcall [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.942604] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 902.942604] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-245f1c1e-aa0f-4b8f-9654-87738a803db9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.979021] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.979021] env[61986]: value = "task-1159903" [ 902.979021] env[61986]: _type = "Task" [ 902.979021] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.988737] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159903, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.041960] env[61986]: DEBUG nova.compute.manager [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.043030] env[61986]: DEBUG nova.compute.manager [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing instance network info cache due to event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 903.043030] env[61986]: DEBUG oslo_concurrency.lockutils [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] Acquiring lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.122465] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159902, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.197670] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159900, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.263802] env[61986]: DEBUG oslo_concurrency.lockutils [req-f845a816-ad8e-44f0-93f5-a5b3c741fd89 req-f73a4470-6232-4819-973d-7f09adfaeab0 service nova] Releasing lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.263802] env[61986]: DEBUG oslo_concurrency.lockutils [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] Acquired lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.264216] env[61986]: DEBUG nova.network.neutron [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 903.287992] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625363} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.287992] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee/f42e92ef-8a90-473c-8662-57cbb9e3f4ee.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 903.287992] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.288410] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac923a58-555d-4bf0-b923-26b412ee63d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.298614] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 903.298614] env[61986]: value = "task-1159904" [ 903.298614] env[61986]: _type = "Task" [ 903.298614] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.311378] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159904, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.496989] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159903, 'name': CreateVM_Task, 'duration_secs': 0.494208} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.497246] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.498050] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.498252] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.498659] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.498972] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb43ae33-8838-43f7-abba-a84afdfd605e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.510144] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 903.510144] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524faa8e-02cc-d3b3-996d-1ed7ceccf0c5" [ 903.510144] env[61986]: _type = "Task" [ 903.510144] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.530918] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524faa8e-02cc-d3b3-996d-1ed7ceccf0c5, 'name': SearchDatastore_Task, 'duration_secs': 0.012381} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.532205] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.532205] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.532587] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.532899] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.533305] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.533890] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6867174e-16b7-4a34-b594-9a47f7babc7d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.549035] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.549035] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.550411] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21db805c-af2e-455e-b9bf-ac66952231d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.556850] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 903.556850] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e78d75-1807-f9e7-d161-0909c19fc81c" [ 903.556850] env[61986]: _type = "Task" [ 903.556850] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.563558] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ca55ae-c7a7-40f7-af67-6fa7569edaa7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.570044] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e78d75-1807-f9e7-d161-0909c19fc81c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.575522] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ee9565-577d-4070-af1b-b1fd9c62ccf2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.614235] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a431327-a2b9-4215-ae09-06287c0bc056 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.625201] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159902, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.629207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963b5068-da27-4fa7-a31f-c5f9fe4041bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.650269] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.689123] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159900, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.810687] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159904, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096269} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.816458] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.816458] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33363fbf-8bc0-4313-9931-209aa9de7e63 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.840318] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee/f42e92ef-8a90-473c-8662-57cbb9e3f4ee.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.840519] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-885ec64b-0de8-4f1e-a04a-8ac9367064fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.864560] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 903.864560] env[61986]: value = "task-1159905" [ 903.864560] env[61986]: _type = "Task" [ 903.864560] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.875765] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159905, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.029316] env[61986]: DEBUG nova.network.neutron [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updated VIF entry in instance network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 904.029729] env[61986]: DEBUG nova.network.neutron [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.071030] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e78d75-1807-f9e7-d161-0909c19fc81c, 'name': SearchDatastore_Task, 'duration_secs': 0.013046} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.071030] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-792aaf5a-7af2-45e6-8604-6c246f5b4703 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.076157] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 904.076157] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cdd91f-e012-19a5-7448-9f244ff08128" [ 904.076157] env[61986]: _type = "Task" [ 904.076157] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.086681] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cdd91f-e012-19a5-7448-9f244ff08128, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.120753] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159902, 'name': Rename_Task, 'duration_secs': 1.054493} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.121709] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 904.121709] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f28410ee-0fda-430e-b529-bc5d5257c05d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.128786] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 904.128786] env[61986]: value = "task-1159906" [ 904.128786] env[61986]: _type = "Task" [ 904.128786] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.138255] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.154532] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.192568] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159900, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.375658] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159905, 'name': ReconfigVM_Task, 'duration_secs': 0.504674} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.376469] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Reconfigured VM instance instance-00000046 to attach disk [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee/f42e92ef-8a90-473c-8662-57cbb9e3f4ee.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.379024] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46865afd-9746-430a-bc02-dfe2fcc20073 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.385270] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 904.385270] env[61986]: value = "task-1159907" [ 904.385270] env[61986]: _type = "Task" [ 904.385270] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.394806] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159907, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.533432] env[61986]: DEBUG oslo_concurrency.lockutils [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] Releasing lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.536218] env[61986]: DEBUG nova.compute.manager [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Received event network-changed-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.536218] env[61986]: DEBUG nova.compute.manager [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Refreshing instance network info cache due to event network-changed-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.536218] env[61986]: DEBUG oslo_concurrency.lockutils [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] Acquiring lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.536218] env[61986]: DEBUG oslo_concurrency.lockutils [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] Acquired lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.536218] env[61986]: DEBUG nova.network.neutron [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Refreshing network info cache for port e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.586971] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cdd91f-e012-19a5-7448-9f244ff08128, 'name': SearchDatastore_Task, 'duration_secs': 0.022738} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.587337] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.587609] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3512cd72-6666-4810-828e-50230956c4b1/3512cd72-6666-4810-828e-50230956c4b1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.587885] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f76de045-f3b4-45f7-a583-1b754e3877d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.595989] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 904.595989] env[61986]: value = "task-1159908" [ 904.595989] env[61986]: _type = "Task" [ 904.595989] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.606870] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159908, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.640246] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159906, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.660196] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 904.660196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.899s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.660360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.798s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.660773] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.662694] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.941s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.662913] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.664721] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.785s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.665730] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.666612] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.335s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.670113] env[61986]: INFO nova.compute.claims [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.693100] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159900, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.425121} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.693100] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/486535fc-ed79-4f4f-a6fb-1cb442ccd708/486535fc-ed79-4f4f-a6fb-1cb442ccd708.vmdk to [datastore2] 6f9349b7-886a-4077-8f6d-a9800ab353fe/6f9349b7-886a-4077-8f6d-a9800ab353fe.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 904.693814] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33f11a0-3a99-4859-ab39-fdfd63e20ce5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.723493] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 6f9349b7-886a-4077-8f6d-a9800ab353fe/6f9349b7-886a-4077-8f6d-a9800ab353fe.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.724705] env[61986]: INFO nova.scheduler.client.report [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleted allocations for instance 46b18b6c-6e40-45b7-9d3f-6177b08a52a5 [ 904.726860] env[61986]: INFO nova.scheduler.client.report [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted allocations for instance 977799b4-2793-4513-9447-483146fc7ac4 [ 904.728681] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd364cde-4050-4ced-afbd-ec21232b3331 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.745519] env[61986]: DEBUG nova.compute.manager [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Received event network-changed-73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.745745] env[61986]: DEBUG nova.compute.manager [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Refreshing instance network info cache due to event network-changed-73b7b19d-3540-46ec-b022-6bc7aaf30b1d. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.745979] env[61986]: DEBUG oslo_concurrency.lockutils [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] Acquiring lock "refresh_cache-3512cd72-6666-4810-828e-50230956c4b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.746148] env[61986]: DEBUG oslo_concurrency.lockutils [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] Acquired lock "refresh_cache-3512cd72-6666-4810-828e-50230956c4b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.746325] env[61986]: DEBUG nova.network.neutron [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Refreshing network info cache for port 73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.748403] env[61986]: INFO nova.scheduler.client.report [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Deleted allocations for instance 40de5f22-0de2-466a-91ab-dcb6ec586dad [ 904.762340] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 904.762340] env[61986]: value = "task-1159909" [ 904.762340] env[61986]: _type = "Task" [ 904.762340] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.775410] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159909, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.797042] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.797164] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.901340] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159907, 'name': Rename_Task, 'duration_secs': 0.235562} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.901731] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 904.901914] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68da5dfd-b137-410f-b66b-068d45b82f78 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.911676] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 904.911676] env[61986]: value = "task-1159910" [ 904.911676] env[61986]: _type = "Task" [ 904.911676] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.921887] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159910, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.108743] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159908, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.139093] env[61986]: DEBUG oslo_vmware.api [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159906, 'name': PowerOnVM_Task, 'duration_secs': 0.91283} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.141641] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 905.141877] env[61986]: INFO nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Took 10.64 seconds to spawn the instance on the hypervisor. [ 905.142080] env[61986]: DEBUG nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.142971] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6ccf0a-966e-4841-87db-6ac2f4248f7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.263290] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6d4071dc-7d4a-427b-82c5-50fcf224bc9e tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "46b18b6c-6e40-45b7-9d3f-6177b08a52a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.471s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.268410] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7862472c-f804-41b0-bd6f-5b3f9ecb68b7 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "977799b4-2793-4513-9447-483146fc7ac4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.015s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.275542] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159909, 'name': ReconfigVM_Task, 'duration_secs': 0.462323} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.275690] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 6f9349b7-886a-4077-8f6d-a9800ab353fe/6f9349b7-886a-4077-8f6d-a9800ab353fe.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.278877] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae7b405-88ed-46f5-8ce0-b40a14c1efbb tempest-MigrationsAdminTest-1193468496 tempest-MigrationsAdminTest-1193468496-project-member] Lock "40de5f22-0de2-466a-91ab-dcb6ec586dad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.456s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.279796] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af5baaef-50a8-4471-adb1-02a24e10ef65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.283914] env[61986]: DEBUG nova.network.neutron [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updated VIF entry in instance network info cache for port e62bf9a8-ba15-4ba9-8409-812d0e3c98a0. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.284293] env[61986]: DEBUG nova.network.neutron [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [{"id": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "address": "fa:16:3e:98:be:f5", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape62bf9a8-ba", "ovs_interfaceid": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.291316] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 905.291316] env[61986]: value = "task-1159911" [ 905.291316] env[61986]: _type = "Task" [ 905.291316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.301752] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159911, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.305139] env[61986]: DEBUG nova.compute.utils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.424913] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159910, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.546184] env[61986]: DEBUG nova.compute.manager [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Received event network-changed-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.546379] env[61986]: DEBUG nova.compute.manager [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Refreshing instance network info cache due to event network-changed-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 905.546599] env[61986]: DEBUG oslo_concurrency.lockutils [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] Acquiring lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.560756] env[61986]: DEBUG nova.network.neutron [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Updated VIF entry in instance network info cache for port 73b7b19d-3540-46ec-b022-6bc7aaf30b1d. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.561646] env[61986]: DEBUG nova.network.neutron [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Updating instance_info_cache with network_info: [{"id": "73b7b19d-3540-46ec-b022-6bc7aaf30b1d", "address": "fa:16:3e:df:30:83", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73b7b19d-35", "ovs_interfaceid": "73b7b19d-3540-46ec-b022-6bc7aaf30b1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.607243] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159908, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55101} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.607525] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3512cd72-6666-4810-828e-50230956c4b1/3512cd72-6666-4810-828e-50230956c4b1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 905.607739] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.608312] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23a329d7-c707-4b20-b34f-4b72d13e434f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.614298] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 905.614298] env[61986]: value = "task-1159912" [ 905.614298] env[61986]: _type = "Task" [ 905.614298] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.622745] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.623017] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.623324] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.623515] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 905.623654] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 905.660389] env[61986]: INFO nova.compute.manager [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Took 35.38 seconds to build instance. [ 905.731474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.731745] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.732010] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.732209] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.732382] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.734730] env[61986]: INFO nova.compute.manager [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Terminating instance [ 905.736906] env[61986]: DEBUG nova.compute.manager [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.737124] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.737989] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4ba32c-429d-43a5-9129-85d84e4fd302 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.749131] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.749131] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac74c656-5872-4eae-820e-a4f788ede397 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.762098] env[61986]: DEBUG oslo_vmware.api [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 905.762098] env[61986]: value = "task-1159913" [ 905.762098] env[61986]: _type = "Task" [ 905.762098] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.770754] env[61986]: DEBUG oslo_vmware.api [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.787414] env[61986]: DEBUG oslo_concurrency.lockutils [req-15801b5e-7904-44c0-833a-ef986bdf402a req-9fb5a9f0-6ddb-498b-8aeb-e21bce02ac25 service nova] Releasing lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.787414] env[61986]: DEBUG oslo_concurrency.lockutils [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] Acquired lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.787574] env[61986]: DEBUG nova.network.neutron [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Refreshing network info cache for port e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.804026] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159911, 'name': Rename_Task, 'duration_secs': 0.173251} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.804426] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.804770] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52e0c55f-76a4-4d79-ac87-24a76f199130 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.808283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.812292] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 905.812292] env[61986]: value = "task-1159914" [ 905.812292] env[61986]: _type = "Task" [ 905.812292] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.824492] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.924505] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159910, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.065152] env[61986]: DEBUG oslo_concurrency.lockutils [req-5c47ac9a-5e47-40b3-ac67-7d23ba8b9d15 req-51dec7eb-1813-4015-9ad1-af1d3be926ad service nova] Releasing lock "refresh_cache-3512cd72-6666-4810-828e-50230956c4b1" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.128370] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065244} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.128838] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.129998] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fd7705-3edd-484f-9afd-18ddd9931b97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.155954] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 3512cd72-6666-4810-828e-50230956c4b1/3512cd72-6666-4810-828e-50230956c4b1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.157319] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fba0bbd0-8ae5-43a7-87a3-957871eef949 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.172171] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22d6c6d-8559-433d-8cc6-baa79d712c74 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.175073] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2796263f-4d43-478c-8116-256459b4d4f4 tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.390s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.183152] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6020ae3d-abdc-41bd-b81c-ca99b8d79a4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.187279] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 906.187279] env[61986]: value = "task-1159915" [ 906.187279] env[61986]: _type = "Task" [ 906.187279] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.220538] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.223370] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d5b614-8a29-4604-a368-9801912c6e45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.230127] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.236206] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84b91a9-de10-4496-b8b6-98ac0f993856 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.252354] env[61986]: DEBUG nova.compute.provider_tree [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.269553] env[61986]: DEBUG oslo_vmware.api [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159913, 'name': PowerOffVM_Task, 'duration_secs': 0.216566} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.269882] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.270161] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.270440] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b45fed3a-7c46-470c-9230-8b161f5c9438 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.323406] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159914, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.342303] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.342577] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.342965] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Deleting the datastore file [datastore2] a101dc3c-ca6b-4a72-a9b4-051b077a10fd {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.343351] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b307acc-90b4-4ade-bb1d-ae0ff1f2494e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.352407] env[61986]: DEBUG oslo_vmware.api [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 906.352407] env[61986]: value = "task-1159917" [ 906.352407] env[61986]: _type = "Task" [ 906.352407] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.361631] env[61986]: DEBUG oslo_vmware.api [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.437473] env[61986]: DEBUG oslo_vmware.api [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1159910, 'name': PowerOnVM_Task, 'duration_secs': 1.199503} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.438667] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 906.439267] env[61986]: INFO nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Took 9.29 seconds to spawn the instance on the hypervisor. [ 906.439267] env[61986]: DEBUG nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.440218] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960547a5-d220-4da5-99a6-8ffb3d4a4c48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.678727] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.708321] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159915, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.723404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.755509] env[61986]: DEBUG nova.scheduler.client.report [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.823515] env[61986]: DEBUG oslo_vmware.api [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159914, 'name': PowerOnVM_Task, 'duration_secs': 0.950244} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.824108] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 906.824514] env[61986]: INFO nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Took 17.06 seconds to spawn the instance on the hypervisor. [ 906.824744] env[61986]: DEBUG nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.825551] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77309de2-f7ee-42cd-bb66-0bcf003b5f73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.864264] env[61986]: DEBUG oslo_vmware.api [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1159917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307286} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.864264] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.864264] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.864264] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.864264] env[61986]: INFO nova.compute.manager [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 906.864264] env[61986]: DEBUG oslo.service.loopingcall [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.864264] env[61986]: DEBUG nova.compute.manager [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.864264] env[61986]: DEBUG nova.network.neutron [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 906.934083] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.934083] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.934083] env[61986]: INFO nova.compute.manager [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Attaching volume 753f1b3b-c85e-4cf3-83c8-a10d93fd03fd to /dev/sdb [ 906.959766] env[61986]: DEBUG nova.network.neutron [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updated VIF entry in instance network info cache for port e62bf9a8-ba15-4ba9-8409-812d0e3c98a0. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.960134] env[61986]: DEBUG nova.network.neutron [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [{"id": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "address": "fa:16:3e:98:be:f5", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape62bf9a8-ba", "ovs_interfaceid": "e62bf9a8-ba15-4ba9-8409-812d0e3c98a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.966906] env[61986]: INFO nova.compute.manager [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Took 35.23 seconds to build instance. [ 906.979235] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89d1880-0d2d-4039-834b-9ea9a5553f60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.989186] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b25c76-c99f-4794-84ed-dbdf66927971 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.005343] env[61986]: DEBUG nova.virt.block_device [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating existing volume attachment record: 75da4169-b023-45b2-8fef-a441ec8814f4 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 907.199373] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159915, 'name': ReconfigVM_Task, 'duration_secs': 0.521626} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.199489] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 3512cd72-6666-4810-828e-50230956c4b1/3512cd72-6666-4810-828e-50230956c4b1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.200162] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab9dd305-a518-429b-b158-3ab5c86f933d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.206895] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 907.206895] env[61986]: value = "task-1159919" [ 907.206895] env[61986]: _type = "Task" [ 907.206895] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.215853] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159919, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.216796] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.261263] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.261860] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.264778] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.549s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.266183] env[61986]: INFO nova.compute.claims [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.349922] env[61986]: INFO nova.compute.manager [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Took 41.87 seconds to build instance. [ 907.468406] env[61986]: DEBUG oslo_concurrency.lockutils [req-3e2574f2-54cb-4adf-b29a-e2574824225c req-bdb90dd4-02aa-4a64-8d0f-5201774f834f service nova] Releasing lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.469177] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bf915db-f2df-4179-80b4-e5a1d345880b tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.699s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.673947] env[61986]: DEBUG nova.network.neutron [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.719960] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159919, 'name': Rename_Task, 'duration_secs': 0.210943} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.720355] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 907.721145] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cfc8f3d-325a-4777-935b-fe462d2d4b77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.728370] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 907.728370] env[61986]: value = "task-1159922" [ 907.728370] env[61986]: _type = "Task" [ 907.728370] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.739426] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.771265] env[61986]: DEBUG nova.compute.utils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.775266] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 907.775474] env[61986]: DEBUG nova.network.neutron [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 907.841406] env[61986]: DEBUG nova.policy [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0567dd93819645f999d224e768614139', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a846059e912047c6bb91b16e4def04de', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 907.852507] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b531c993-d550-4b84-84fe-44dde2d60caa tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.286s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.874677] env[61986]: DEBUG nova.compute.manager [req-74ef7e5e-ce11-4893-801b-cddb8dfbc1ec req-cb84c403-8622-4a9a-a4e7-5d28aa00bf38 service nova] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Received event network-vif-deleted-e62bf9a8-ba15-4ba9-8409-812d0e3c98a0 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.970435] env[61986]: DEBUG nova.compute.manager [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.971253] env[61986]: DEBUG nova.compute.manager [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing instance network info cache due to event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 907.971759] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] Acquiring lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.972131] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] Acquired lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.972393] env[61986]: DEBUG nova.network.neutron [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.976516] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.183698] env[61986]: INFO nova.compute.manager [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Took 1.32 seconds to deallocate network for instance. [ 908.200782] env[61986]: DEBUG nova.network.neutron [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Successfully created port: 7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.241624] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159922, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.274313] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.355936] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.505061] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.691743] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.742141] env[61986]: DEBUG oslo_vmware.api [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159922, 'name': PowerOnVM_Task, 'duration_secs': 0.769252} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.742536] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.742775] env[61986]: INFO nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Took 9.43 seconds to spawn the instance on the hypervisor. [ 908.742981] env[61986]: DEBUG nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.745184] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a7964b-da99-48d3-8a78-6b1292bf91fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.800437] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4318753b-c88b-410a-9e2f-b6892d0c9d46 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.809522] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d88047-43ef-4c97-808e-a52c6debc5cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.852087] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a876f6ae-3841-4365-a4bd-2cbf975d2d2d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.861811] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6dfee7-26cd-4852-be88-a9f52f1339a1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.883882] env[61986]: DEBUG nova.compute.provider_tree [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.885957] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.920638] env[61986]: DEBUG nova.network.neutron [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updated VIF entry in instance network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 908.921057] env[61986]: DEBUG nova.network.neutron [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.278556] env[61986]: INFO nova.compute.manager [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Took 33.30 seconds to build instance. [ 909.285318] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.313434] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.313779] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.314150] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.321021] env[61986]: DEBUG nova.virt.hardware [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.321021] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb08f6a3-8bca-4497-b942-d065b4e42a87 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.330176] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b40011-3151-478a-881c-0bc81a5dcd6e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.390454] env[61986]: DEBUG nova.scheduler.client.report [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.424381] env[61986]: DEBUG oslo_concurrency.lockutils [req-3ce4e786-f7fa-4109-9c93-5f9cb24aba5d req-424eb98a-6bae-4ac5-916c-188a3687ca67 service nova] Releasing lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.782996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "6f9349b7-886a-4077-8f6d-a9800ab353fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.782996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.782996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "6f9349b7-886a-4077-8f6d-a9800ab353fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.782996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.782996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.784226] env[61986]: INFO nova.compute.manager [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Terminating instance [ 909.786234] env[61986]: DEBUG nova.compute.manager [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.786557] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 909.787248] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bc156c37-875f-4f03-a6d6-fa837992a955 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "3512cd72-6666-4810-828e-50230956c4b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.371s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.788666] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccbfd76-6d8c-471a-bc6b-25fbb34d4b53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.797295] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.797719] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9af3cfe6-8a02-47e6-a013-4d8bf03e77dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.803997] env[61986]: DEBUG oslo_vmware.api [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 909.803997] env[61986]: value = "task-1159924" [ 909.803997] env[61986]: _type = "Task" [ 909.803997] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.813912] env[61986]: DEBUG oslo_vmware.api [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.894551] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.629s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.897597] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.904555] env[61986]: DEBUG nova.network.neutron [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Successfully updated port: 7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.905349] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.188s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.907122] env[61986]: INFO nova.compute.claims [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.087415] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c0056b-3bd5-91e9-a229-66eefcd769a1/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 910.088563] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b59203-ee35-4ab6-b1ec-dde89a359fdf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.095394] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c0056b-3bd5-91e9-a229-66eefcd769a1/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 910.095545] env[61986]: ERROR oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c0056b-3bd5-91e9-a229-66eefcd769a1/disk-0.vmdk due to incomplete transfer. [ 910.095796] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9bd0560c-9c0e-4335-99b5-13a953ca1fef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.102538] env[61986]: DEBUG oslo_vmware.rw_handles [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c0056b-3bd5-91e9-a229-66eefcd769a1/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 910.102742] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Uploaded image f56ca176-82f3-4ea2-8aaa-08ad8bc1fba2 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 910.105158] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 910.105684] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-89e0a7b7-0a54-40a7-865e-e590b0894361 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.111972] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 910.111972] env[61986]: value = "task-1159925" [ 910.111972] env[61986]: _type = "Task" [ 910.111972] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.120335] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159925, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.176418] env[61986]: DEBUG nova.compute.manager [req-f23809c5-6407-403a-9f0c-cd146a049a62 req-6a3111a9-5d59-4f86-8857-5ee1ea732e64 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Received event network-vif-plugged-7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.176681] env[61986]: DEBUG oslo_concurrency.lockutils [req-f23809c5-6407-403a-9f0c-cd146a049a62 req-6a3111a9-5d59-4f86-8857-5ee1ea732e64 service nova] Acquiring lock "59cfee64-ef79-4b8f-a703-c8812551d12f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.176869] env[61986]: DEBUG oslo_concurrency.lockutils [req-f23809c5-6407-403a-9f0c-cd146a049a62 req-6a3111a9-5d59-4f86-8857-5ee1ea732e64 service nova] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.177061] env[61986]: DEBUG oslo_concurrency.lockutils [req-f23809c5-6407-403a-9f0c-cd146a049a62 req-6a3111a9-5d59-4f86-8857-5ee1ea732e64 service nova] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.177240] env[61986]: DEBUG nova.compute.manager [req-f23809c5-6407-403a-9f0c-cd146a049a62 req-6a3111a9-5d59-4f86-8857-5ee1ea732e64 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] No waiting events found dispatching network-vif-plugged-7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.177410] env[61986]: WARNING nova.compute.manager [req-f23809c5-6407-403a-9f0c-cd146a049a62 req-6a3111a9-5d59-4f86-8857-5ee1ea732e64 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Received unexpected event network-vif-plugged-7b26e9ba-74ad-493c-8508-422bdb97e8db for instance with vm_state building and task_state spawning. [ 910.179762] env[61986]: DEBUG nova.compute.manager [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.179947] env[61986]: DEBUG nova.compute.manager [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing instance network info cache due to event network-changed-812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 910.180161] env[61986]: DEBUG oslo_concurrency.lockutils [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] Acquiring lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.180305] env[61986]: DEBUG oslo_concurrency.lockutils [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] Acquired lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.180517] env[61986]: DEBUG nova.network.neutron [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Refreshing network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.181790] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "b833adc9-2967-4b0b-81c2-0b8deac20f69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.181997] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.182213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "b833adc9-2967-4b0b-81c2-0b8deac20f69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.182393] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.182556] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.184742] env[61986]: INFO nova.compute.manager [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Terminating instance [ 910.186709] env[61986]: DEBUG nova.compute.manager [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.186709] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.187407] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff2a6c7-2950-4ec8-8240-40a9d18545bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.198013] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.198277] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df5993d7-887f-477b-a6f5-20222099c56a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.204224] env[61986]: DEBUG oslo_vmware.api [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 910.204224] env[61986]: value = "task-1159926" [ 910.204224] env[61986]: _type = "Task" [ 910.204224] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.212560] env[61986]: DEBUG oslo_vmware.api [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.252656] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.317125] env[61986]: DEBUG oslo_vmware.api [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159924, 'name': PowerOffVM_Task, 'duration_secs': 0.188273} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.318219] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.318219] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.318219] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c685a279-bdda-4e2b-9f70-de2d47cf7cb8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.323058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "3512cd72-6666-4810-828e-50230956c4b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.323058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "3512cd72-6666-4810-828e-50230956c4b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.323058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "3512cd72-6666-4810-828e-50230956c4b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.323058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "3512cd72-6666-4810-828e-50230956c4b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.323058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "3512cd72-6666-4810-828e-50230956c4b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.324052] env[61986]: INFO nova.compute.manager [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Terminating instance [ 910.326131] env[61986]: DEBUG nova.compute.manager [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.326511] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.327668] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2dba9f0-4697-4391-9799-f184338ab185 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.337638] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.338093] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37b1f8c4-11df-433e-a529-4b7b557cd5fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.347152] env[61986]: DEBUG oslo_vmware.api [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 910.347152] env[61986]: value = "task-1159928" [ 910.347152] env[61986]: _type = "Task" [ 910.347152] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.357895] env[61986]: DEBUG oslo_vmware.api [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159928, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.387456] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.387783] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.388041] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleting the datastore file [datastore2] 6f9349b7-886a-4077-8f6d-a9800ab353fe {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.388400] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b44bf5d3-1ee8-461e-a188-00b74527eb0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.395409] env[61986]: DEBUG oslo_vmware.api [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 910.395409] env[61986]: value = "task-1159929" [ 910.395409] env[61986]: _type = "Task" [ 910.395409] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.408861] env[61986]: DEBUG nova.compute.utils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.410598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "refresh_cache-59cfee64-ef79-4b8f-a703-c8812551d12f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.410863] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired lock "refresh_cache-59cfee64-ef79-4b8f-a703-c8812551d12f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.411045] env[61986]: DEBUG nova.network.neutron [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.412079] env[61986]: DEBUG oslo_vmware.api [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.416346] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.418139] env[61986]: DEBUG nova.network.neutron [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 910.473559] env[61986]: DEBUG nova.policy [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e34c464681464270a5ae8c02b6e3fb75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ea5fc77300041f5aeccb8773300ef0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 910.624262] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159925, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.717423] env[61986]: DEBUG oslo_vmware.api [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159926, 'name': PowerOffVM_Task, 'duration_secs': 0.276491} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.720541] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.720744] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.721083] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d4ae042-3898-465e-a017-6874292334b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.790099] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.790333] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.790518] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Deleting the datastore file [datastore2] b833adc9-2967-4b0b-81c2-0b8deac20f69 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.790779] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62948dbc-481e-43f9-a406-064c99cc0c97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.799346] env[61986]: DEBUG oslo_vmware.api [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for the task: (returnval){ [ 910.799346] env[61986]: value = "task-1159931" [ 910.799346] env[61986]: _type = "Task" [ 910.799346] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.807811] env[61986]: DEBUG oslo_vmware.api [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.856017] env[61986]: DEBUG oslo_vmware.api [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159928, 'name': PowerOffVM_Task, 'duration_secs': 0.205861} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.856763] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 910.856939] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 910.857317] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8587cc42-6785-42dd-b929-1596a81aeea8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.914833] env[61986]: DEBUG oslo_vmware.api [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1159929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182224} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.915747] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 910.918770] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.918770] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 910.918770] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 910.918952] env[61986]: INFO nova.compute.manager [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Took 1.13 seconds to destroy the instance on the hypervisor. [ 910.924020] env[61986]: DEBUG oslo.service.loopingcall [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.924242] env[61986]: DEBUG nova.compute.manager [-] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.924353] env[61986]: DEBUG nova.network.neutron [-] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 910.932283] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 910.932283] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 910.932283] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleting the datastore file [datastore1] 3512cd72-6666-4810-828e-50230956c4b1 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.932671] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-497cb72c-b1eb-4d8a-8429-56f367f93aba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.939664] env[61986]: DEBUG oslo_vmware.api [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 910.939664] env[61986]: value = "task-1159933" [ 910.939664] env[61986]: _type = "Task" [ 910.939664] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.954176] env[61986]: DEBUG oslo_vmware.api [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.976420] env[61986]: DEBUG nova.network.neutron [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Successfully created port: 495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.020753] env[61986]: DEBUG nova.network.neutron [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 911.077360] env[61986]: DEBUG nova.network.neutron [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updated VIF entry in instance network info cache for port 812b5c34-903f-47de-b3f8-649b3a84eb27. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.077735] env[61986]: DEBUG nova.network.neutron [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [{"id": "812b5c34-903f-47de-b3f8-649b3a84eb27", "address": "fa:16:3e:2c:c4:00", "network": {"id": "e0ef3efd-a586-4176-95a1-5952867f1236", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1137193111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "50561f3dfae840348c86bcce568d354f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap812b5c34-90", "ovs_interfaceid": "812b5c34-903f-47de-b3f8-649b3a84eb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.125366] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159925, 'name': Destroy_Task, 'duration_secs': 0.786843} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.125798] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroyed the VM [ 911.126164] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 911.126667] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e8d7cff1-2617-4a5b-98a3-6c15b877a72a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.136982] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 911.136982] env[61986]: value = "task-1159934" [ 911.136982] env[61986]: _type = "Task" [ 911.136982] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.149413] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159934, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.310946] env[61986]: DEBUG oslo_vmware.api [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Task: {'id': task-1159931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16018} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.311282] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.311404] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.311586] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.311760] env[61986]: INFO nova.compute.manager [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Took 1.13 seconds to destroy the instance on the hypervisor. [ 911.312009] env[61986]: DEBUG oslo.service.loopingcall [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.312518] env[61986]: DEBUG nova.compute.manager [-] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.312518] env[61986]: DEBUG nova.network.neutron [-] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.422795] env[61986]: DEBUG nova.network.neutron [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Updating instance_info_cache with network_info: [{"id": "7b26e9ba-74ad-493c-8508-422bdb97e8db", "address": "fa:16:3e:1c:b9:f3", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.151", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b26e9ba-74", "ovs_interfaceid": "7b26e9ba-74ad-493c-8508-422bdb97e8db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.458602] env[61986]: DEBUG oslo_vmware.api [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1159933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160301} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.460227] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.460227] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.460227] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.460227] env[61986]: INFO nova.compute.manager [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Took 1.13 seconds to destroy the instance on the hypervisor. [ 911.460227] env[61986]: DEBUG oslo.service.loopingcall [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.461116] env[61986]: DEBUG nova.compute.manager [-] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.461116] env[61986]: DEBUG nova.network.neutron [-] [instance: 3512cd72-6666-4810-828e-50230956c4b1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.505545] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128b9ba6-37ba-48ec-be6d-64c927c7b835 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.513027] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168aee67-9926-4db9-a901-32a26e8f07df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.547920] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d186bf-f7e3-4887-bebb-3042119a5991 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.553572] env[61986]: DEBUG nova.compute.manager [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Stashing vm_state: active {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 911.559883] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 911.560083] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252383', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'name': 'volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'serial': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 911.561473] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13be410f-400c-4fa9-b177-ec7390a88f54 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.566063] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de42c24-ae76-4838-ac4a-3452bf8e02ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.580582] env[61986]: DEBUG nova.compute.provider_tree [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.592132] env[61986]: DEBUG oslo_concurrency.lockutils [req-0551c073-473d-4ef8-8861-97739008e70e req-a2b7901a-0014-43e6-b3b4-7ae2ae3d4332 service nova] Releasing lock "refresh_cache-86c9e45c-b3f1-4004-9acc-190b11a4a926" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.593990] env[61986]: DEBUG nova.scheduler.client.report [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.597510] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8f83c4-14d5-4656-a519-bfd202bbc2cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.624712] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd/volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.629019] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6806b2d1-62ea-4007-aac9-3f65363d3c23 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.644229] env[61986]: DEBUG nova.network.neutron [-] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.655594] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159934, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.657089] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 911.657089] env[61986]: value = "task-1159935" [ 911.657089] env[61986]: _type = "Task" [ 911.657089] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.666286] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159935, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.755604] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.755826] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.925443] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Releasing lock "refresh_cache-59cfee64-ef79-4b8f-a703-c8812551d12f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.925777] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Instance network_info: |[{"id": "7b26e9ba-74ad-493c-8508-422bdb97e8db", "address": "fa:16:3e:1c:b9:f3", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.151", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b26e9ba-74", "ovs_interfaceid": "7b26e9ba-74ad-493c-8508-422bdb97e8db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 911.926246] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:b9:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b26e9ba-74ad-493c-8508-422bdb97e8db', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.934356] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating folder: Project (a846059e912047c6bb91b16e4def04de). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 911.935432] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.937876] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c721661f-5b09-49a1-8d33-e103492548d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.950092] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Created folder: Project (a846059e912047c6bb91b16e4def04de) in parent group-v252271. [ 911.950367] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating folder: Instances. Parent ref: group-v252384. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 911.950695] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5023464-2fef-4752-889b-88ebd77ce715 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.960487] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Created folder: Instances in parent group-v252384. [ 911.960736] env[61986]: DEBUG oslo.service.loopingcall [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.960932] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 911.961538] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dffcd9a8-05d7-4ace-87cf-ea745cf6ea67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.984572] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.984572] env[61986]: value = "task-1159938" [ 911.984572] env[61986]: _type = "Task" [ 911.984572] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.989123] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.989397] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.989578] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.989776] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.989920] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.990089] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.990306] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.990461] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.990617] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.990776] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.991042] env[61986]: DEBUG nova.virt.hardware [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.991886] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c834a5c-2636-4a3e-8169-7f33987611b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.009706] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159938, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.009706] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bfe480-3868-4c2d-9a13-8ac09d0c3f27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.074145] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.102183] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.102572] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.106948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.247s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.108630] env[61986]: INFO nova.compute.claims [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.147728] env[61986]: INFO nova.compute.manager [-] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Took 1.22 seconds to deallocate network for instance. [ 912.163109] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159934, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.175413] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.261951] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.325954] env[61986]: DEBUG nova.network.neutron [-] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.354018] env[61986]: DEBUG nova.compute.manager [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Received event network-changed-7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.354390] env[61986]: DEBUG nova.compute.manager [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Refreshing instance network info cache due to event network-changed-7b26e9ba-74ad-493c-8508-422bdb97e8db. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 912.354390] env[61986]: DEBUG oslo_concurrency.lockutils [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] Acquiring lock "refresh_cache-59cfee64-ef79-4b8f-a703-c8812551d12f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.354538] env[61986]: DEBUG oslo_concurrency.lockutils [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] Acquired lock "refresh_cache-59cfee64-ef79-4b8f-a703-c8812551d12f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.354647] env[61986]: DEBUG nova.network.neutron [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Refreshing network info cache for port 7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.374604] env[61986]: DEBUG nova.compute.manager [req-e6050245-7c1c-4fd8-a4d5-e055f472d09f req-146795c0-7da1-4f7e-a409-b70279cdaa89 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Received event network-vif-deleted-ff9551fd-6a35-45fb-b841-fb3e1c73861c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.374604] env[61986]: INFO nova.compute.manager [req-e6050245-7c1c-4fd8-a4d5-e055f472d09f req-146795c0-7da1-4f7e-a409-b70279cdaa89 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Neutron deleted interface ff9551fd-6a35-45fb-b841-fb3e1c73861c; detaching it from the instance and deleting it from the info cache [ 912.374604] env[61986]: DEBUG nova.network.neutron [req-e6050245-7c1c-4fd8-a4d5-e055f472d09f req-146795c0-7da1-4f7e-a409-b70279cdaa89 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.492877] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159938, 'name': CreateVM_Task, 'duration_secs': 0.337776} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.492877] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.493612] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.493788] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.494137] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.494420] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a3e3631-c6b9-4c54-b1ca-5142be6f9c2c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.500067] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 912.500067] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aa1b6e-da63-07c9-d41f-5e544b9585b5" [ 912.500067] env[61986]: _type = "Task" [ 912.500067] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.513179] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aa1b6e-da63-07c9-d41f-5e544b9585b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.515493] env[61986]: DEBUG nova.network.neutron [-] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.609158] env[61986]: DEBUG nova.compute.utils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.610818] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.611033] env[61986]: DEBUG nova.network.neutron [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.661040] env[61986]: DEBUG oslo_vmware.api [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159934, 'name': RemoveSnapshot_Task, 'duration_secs': 1.136875} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.661040] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 912.661040] env[61986]: INFO nova.compute.manager [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 16.52 seconds to snapshot the instance on the hypervisor. [ 912.663143] env[61986]: WARNING oslo_messaging._drivers.amqpdriver [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 912.664658] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.674332] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159935, 'name': ReconfigVM_Task, 'duration_secs': 0.704948} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.674605] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfigured VM instance instance-0000003a to attach disk [datastore2] volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd/volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.680181] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b5a6c97-af88-4073-bb69-782119539169 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.696393] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 912.696393] env[61986]: value = "task-1159939" [ 912.696393] env[61986]: _type = "Task" [ 912.696393] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.701750] env[61986]: DEBUG nova.policy [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0567dd93819645f999d224e768614139', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a846059e912047c6bb91b16e4def04de', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 912.709889] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159939, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.784297] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.831896] env[61986]: INFO nova.compute.manager [-] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Took 1.37 seconds to deallocate network for instance. [ 912.877633] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9f65781-2c2c-4373-9087-91a36fcf0f1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.887012] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cee66b-6ffe-44c5-8d0f-20e91e7be7b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.906885] env[61986]: DEBUG nova.compute.manager [req-e96d922f-d670-4973-9637-aca20dd1c51e req-9e15148d-f714-4053-8f4a-5525e2f0eee2 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Received event network-vif-plugged-495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 912.907138] env[61986]: DEBUG oslo_concurrency.lockutils [req-e96d922f-d670-4973-9637-aca20dd1c51e req-9e15148d-f714-4053-8f4a-5525e2f0eee2 service nova] Acquiring lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.907434] env[61986]: DEBUG oslo_concurrency.lockutils [req-e96d922f-d670-4973-9637-aca20dd1c51e req-9e15148d-f714-4053-8f4a-5525e2f0eee2 service nova] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.907522] env[61986]: DEBUG oslo_concurrency.lockutils [req-e96d922f-d670-4973-9637-aca20dd1c51e req-9e15148d-f714-4053-8f4a-5525e2f0eee2 service nova] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.907692] env[61986]: DEBUG nova.compute.manager [req-e96d922f-d670-4973-9637-aca20dd1c51e req-9e15148d-f714-4053-8f4a-5525e2f0eee2 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] No waiting events found dispatching network-vif-plugged-495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.907860] env[61986]: WARNING nova.compute.manager [req-e96d922f-d670-4973-9637-aca20dd1c51e req-9e15148d-f714-4053-8f4a-5525e2f0eee2 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Received unexpected event network-vif-plugged-495a4f17-b010-4427-bf0f-e72582cd1dc1 for instance with vm_state building and task_state spawning. [ 912.927629] env[61986]: DEBUG nova.compute.manager [req-e6050245-7c1c-4fd8-a4d5-e055f472d09f req-146795c0-7da1-4f7e-a409-b70279cdaa89 service nova] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Detach interface failed, port_id=ff9551fd-6a35-45fb-b841-fb3e1c73861c, reason: Instance b833adc9-2967-4b0b-81c2-0b8deac20f69 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 913.018230] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "076c8e91-af70-489f-89d5-f598166fafc6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.019030] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "076c8e91-af70-489f-89d5-f598166fafc6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.019802] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "076c8e91-af70-489f-89d5-f598166fafc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.020156] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "076c8e91-af70-489f-89d5-f598166fafc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.020704] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "076c8e91-af70-489f-89d5-f598166fafc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.024401] env[61986]: INFO nova.compute.manager [-] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Took 1.71 seconds to deallocate network for instance. [ 913.024632] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52aa1b6e-da63-07c9-d41f-5e544b9585b5, 'name': SearchDatastore_Task, 'duration_secs': 0.015129} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.025229] env[61986]: INFO nova.compute.manager [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Terminating instance [ 913.029129] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.029248] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.029476] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.029620] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.029836] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.034410] env[61986]: DEBUG nova.compute.manager [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.034833] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 913.035161] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49643e20-6f54-4d6d-a169-f1849346f0e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.038165] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41a10c2-5534-4a69-bb45-e8bbc610f7f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.047540] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.048842] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7c0bb43-2723-4e84-9c6a-85fa028881e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.053235] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.053467] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 913.054976] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-391db2fc-4a1d-40be-b4bf-4fd361abc617 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.059549] env[61986]: DEBUG oslo_vmware.api [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 913.059549] env[61986]: value = "task-1159940" [ 913.059549] env[61986]: _type = "Task" [ 913.059549] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.064974] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 913.064974] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526250e8-2016-28ce-af34-004bf4ae8d8b" [ 913.064974] env[61986]: _type = "Task" [ 913.064974] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.074134] env[61986]: DEBUG oslo_vmware.api [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.080368] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526250e8-2016-28ce-af34-004bf4ae8d8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.101408] env[61986]: DEBUG nova.network.neutron [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Successfully created port: 9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.114480] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.145726] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.147026] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.147026] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.147026] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.147206] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.150834] env[61986]: INFO nova.compute.manager [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Terminating instance [ 913.153401] env[61986]: DEBUG nova.compute.manager [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.153613] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 913.154556] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8070fcdc-6cad-45ba-bb1d-2dda1a3e097c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.163361] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.163635] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98582a10-40ea-4115-82a6-a0edc934a95f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.180482] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 913.180482] env[61986]: value = "task-1159941" [ 913.180482] env[61986]: _type = "Task" [ 913.180482] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.191451] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.208051] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159939, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.249896] env[61986]: DEBUG nova.compute.manager [None req-df649c56-fbeb-454d-8076-02536ac81772 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Found 2 images (rotation: 2) {{(pid=61986) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 913.284200] env[61986]: DEBUG nova.network.neutron [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Updated VIF entry in instance network info cache for port 7b26e9ba-74ad-493c-8508-422bdb97e8db. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.285852] env[61986]: DEBUG nova.network.neutron [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Updating instance_info_cache with network_info: [{"id": "7b26e9ba-74ad-493c-8508-422bdb97e8db", "address": "fa:16:3e:1c:b9:f3", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.151", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b26e9ba-74", "ovs_interfaceid": "7b26e9ba-74ad-493c-8508-422bdb97e8db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.337609] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.535999] env[61986]: DEBUG nova.network.neutron [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Successfully updated port: 495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.545354] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.577889] env[61986]: DEBUG oslo_vmware.api [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159940, 'name': PowerOffVM_Task, 'duration_secs': 0.246953} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.578156] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526250e8-2016-28ce-af34-004bf4ae8d8b, 'name': SearchDatastore_Task, 'duration_secs': 0.016485} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.580759] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.580950] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 913.581919] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c480869-fdb1-4a3f-9dc5-b263e20ba07a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.584196] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b6136f0-26fb-4b85-8ef0-c042f5856108 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.588636] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 913.588636] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a16d98-1b87-7b21-edb7-5dad5f5ddda8" [ 913.588636] env[61986]: _type = "Task" [ 913.588636] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.598950] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a16d98-1b87-7b21-edb7-5dad5f5ddda8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.674043] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 913.674286] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 913.674472] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleting the datastore file [datastore1] 076c8e91-af70-489f-89d5-f598166fafc6 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.674729] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-165be996-7461-47f2-888f-59694a5265fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.682127] env[61986]: DEBUG oslo_vmware.api [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 913.682127] env[61986]: value = "task-1159943" [ 913.682127] env[61986]: _type = "Task" [ 913.682127] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.689718] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdd760c-7bfb-4c53-a5e8-2c8f26fb1f20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.699144] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.707831] env[61986]: DEBUG oslo_vmware.api [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.709716] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fc795b-832a-4ffc-9624-453ced171e27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.716787] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159939, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.743421] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05d0732-dc9b-4277-9ea8-ab06bc2dfb20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.750489] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0386a22e-48f3-41f9-b935-4ac02e524e54 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.765943] env[61986]: DEBUG nova.compute.provider_tree [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.788809] env[61986]: DEBUG oslo_concurrency.lockutils [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] Releasing lock "refresh_cache-59cfee64-ef79-4b8f-a703-c8812551d12f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.789114] env[61986]: DEBUG nova.compute.manager [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Received event network-vif-deleted-ddc2cc86-1e6d-4436-b6f1-5c967d85c128 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.789311] env[61986]: DEBUG nova.compute.manager [req-852e847d-d5cd-45ac-8e19-b0a4c9c2fc33 req-2be5ebd3-4fd5-4eae-b1db-54b8d3aaccea service nova] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Received event network-vif-deleted-73b7b19d-3540-46ec-b022-6bc7aaf30b1d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.038182] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "refresh_cache-49bcac13-6dde-4a28-9131-f9a1e1d2e386" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.038336] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "refresh_cache-49bcac13-6dde-4a28-9131-f9a1e1d2e386" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.038481] env[61986]: DEBUG nova.network.neutron [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.100581] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a16d98-1b87-7b21-edb7-5dad5f5ddda8, 'name': SearchDatastore_Task, 'duration_secs': 0.008702} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.100856] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.101172] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 59cfee64-ef79-4b8f-a703-c8812551d12f/59cfee64-ef79-4b8f-a703-c8812551d12f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 914.101442] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-468ef0b2-a6fb-4254-a5cc-5841bd04569d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.108391] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 914.108391] env[61986]: value = "task-1159944" [ 914.108391] env[61986]: _type = "Task" [ 914.108391] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.116575] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.131342] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.161498] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.161817] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.166156] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.166366] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.166578] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.166830] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.167152] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.167284] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.167503] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.167723] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.168093] env[61986]: DEBUG nova.virt.hardware [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.168988] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14e5b87-0058-4077-bb78-6db4c0a57a83 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.177764] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3a7ced-a7dc-4a1d-a2a3-f6df162da203 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.204743] env[61986]: DEBUG nova.compute.manager [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.204743] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159941, 'name': PowerOffVM_Task, 'duration_secs': 0.850236} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.207251] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51479b4d-1471-4857-ac88-82250cdd05ca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.210377] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 914.210561] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 914.214473] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1f9b63e-7f56-419c-8eec-c61eed4a25cc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.216363] env[61986]: DEBUG oslo_vmware.api [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140919} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.217509] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.217809] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.218091] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 914.218427] env[61986]: INFO nova.compute.manager [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 914.218834] env[61986]: DEBUG oslo.service.loopingcall [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.223431] env[61986]: DEBUG nova.compute.manager [-] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.223509] env[61986]: DEBUG nova.network.neutron [-] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.227031] env[61986]: DEBUG oslo_vmware.api [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159939, 'name': ReconfigVM_Task, 'duration_secs': 1.151927} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.230476] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252383', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'name': 'volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'serial': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 914.270143] env[61986]: DEBUG nova.scheduler.client.report [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.280438] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.280438] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.280650] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleting the datastore file [datastore1] 52eda388-05f4-416c-ac9d-bd1c1e31ba9c {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.280815] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49cd69e9-9d2c-4e38-b6a2-098d411bfd31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.287780] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for the task: (returnval){ [ 914.287780] env[61986]: value = "task-1159946" [ 914.287780] env[61986]: _type = "Task" [ 914.287780] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.296823] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.385811] env[61986]: DEBUG nova.compute.manager [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Received event network-changed-495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.386033] env[61986]: DEBUG nova.compute.manager [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Refreshing instance network info cache due to event network-changed-495a4f17-b010-4427-bf0f-e72582cd1dc1. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 914.386262] env[61986]: DEBUG oslo_concurrency.lockutils [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] Acquiring lock "refresh_cache-49bcac13-6dde-4a28-9131-f9a1e1d2e386" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.593450] env[61986]: DEBUG nova.network.neutron [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 914.619170] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159944, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.729204] env[61986]: INFO nova.compute.manager [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] instance snapshotting [ 914.729968] env[61986]: DEBUG nova.objects.instance [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'flavor' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.775032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.668s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.775568] env[61986]: DEBUG nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.780036] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.740s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.780241] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.782358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.636s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.783962] env[61986]: INFO nova.compute.claims [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.797476] env[61986]: DEBUG oslo_vmware.api [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Task: {'id': task-1159946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285316} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.797724] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.798013] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.798082] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 914.798261] env[61986]: INFO nova.compute.manager [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Took 1.64 seconds to destroy the instance on the hypervisor. [ 914.798500] env[61986]: DEBUG oslo.service.loopingcall [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.798692] env[61986]: DEBUG nova.compute.manager [-] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.798799] env[61986]: DEBUG nova.network.neutron [-] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.828125] env[61986]: INFO nova.scheduler.client.report [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Deleted allocations for instance 4f13b629-e2a7-4668-9d77-eb638078e246 [ 914.866855] env[61986]: DEBUG nova.compute.manager [req-2f92cfac-47a3-4047-8f07-209f4931c33a req-cdb197f9-9782-4522-8e1d-547ffeb4e4c4 service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Received event network-vif-plugged-9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.867480] env[61986]: DEBUG oslo_concurrency.lockutils [req-2f92cfac-47a3-4047-8f07-209f4931c33a req-cdb197f9-9782-4522-8e1d-547ffeb4e4c4 service nova] Acquiring lock "6488f68e-7b84-4462-aef5-25d02db504f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.867700] env[61986]: DEBUG oslo_concurrency.lockutils [req-2f92cfac-47a3-4047-8f07-209f4931c33a req-cdb197f9-9782-4522-8e1d-547ffeb4e4c4 service nova] Lock "6488f68e-7b84-4462-aef5-25d02db504f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.867872] env[61986]: DEBUG oslo_concurrency.lockutils [req-2f92cfac-47a3-4047-8f07-209f4931c33a req-cdb197f9-9782-4522-8e1d-547ffeb4e4c4 service nova] Lock "6488f68e-7b84-4462-aef5-25d02db504f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.868476] env[61986]: DEBUG nova.compute.manager [req-2f92cfac-47a3-4047-8f07-209f4931c33a req-cdb197f9-9782-4522-8e1d-547ffeb4e4c4 service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] No waiting events found dispatching network-vif-plugged-9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.868889] env[61986]: WARNING nova.compute.manager [req-2f92cfac-47a3-4047-8f07-209f4931c33a req-cdb197f9-9782-4522-8e1d-547ffeb4e4c4 service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Received unexpected event network-vif-plugged-9afcb8ca-5b37-41f4-a4de-a97cf229512b for instance with vm_state building and task_state spawning. [ 914.977045] env[61986]: DEBUG nova.network.neutron [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Updating instance_info_cache with network_info: [{"id": "495a4f17-b010-4427-bf0f-e72582cd1dc1", "address": "fa:16:3e:f6:2a:40", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap495a4f17-b0", "ovs_interfaceid": "495a4f17-b010-4427-bf0f-e72582cd1dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.994952] env[61986]: DEBUG nova.network.neutron [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Successfully updated port: 9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.120688] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528566} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.121115] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 59cfee64-ef79-4b8f-a703-c8812551d12f/59cfee64-ef79-4b8f-a703-c8812551d12f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 915.121254] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.121494] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5244aaf-31b0-4c58-b9a9-a2f62550f0e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.127824] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 915.127824] env[61986]: value = "task-1159947" [ 915.127824] env[61986]: _type = "Task" [ 915.127824] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.137699] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.244619] env[61986]: DEBUG nova.network.neutron [-] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.247144] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76c6575-eac0-4b17-b544-945c87f76f72 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.272947] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3b6577-bb54-44c9-83a7-6c90ada4814e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.289394] env[61986]: DEBUG nova.compute.utils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.293416] env[61986]: DEBUG nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 915.294254] env[61986]: DEBUG nova.objects.instance [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lazy-loading 'flavor' on Instance uuid a0fa2cbe-d8c3-462a-9744-aaa36a8d314e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.339408] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c922a5a3-de25-4d71-9e6d-f9ec83981d01 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "4f13b629-e2a7-4668-9d77-eb638078e246" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.742s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.483025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "refresh_cache-49bcac13-6dde-4a28-9131-f9a1e1d2e386" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.483025] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Instance network_info: |[{"id": "495a4f17-b010-4427-bf0f-e72582cd1dc1", "address": "fa:16:3e:f6:2a:40", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap495a4f17-b0", "ovs_interfaceid": "495a4f17-b010-4427-bf0f-e72582cd1dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.483025] env[61986]: DEBUG oslo_concurrency.lockutils [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] Acquired lock "refresh_cache-49bcac13-6dde-4a28-9131-f9a1e1d2e386" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.483025] env[61986]: DEBUG nova.network.neutron [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Refreshing network info cache for port 495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 915.483025] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:2a:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0954fad3-d24d-496c-83e6-a09d3cb556fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '495a4f17-b010-4427-bf0f-e72582cd1dc1', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.498119] env[61986]: DEBUG oslo.service.loopingcall [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.499563] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "refresh_cache-6488f68e-7b84-4462-aef5-25d02db504f6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.499650] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired lock "refresh_cache-6488f68e-7b84-4462-aef5-25d02db504f6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.499848] env[61986]: DEBUG nova.network.neutron [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.501947] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.502447] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3641b4a4-e948-45e3-ab65-efb5275365df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.524061] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.524061] env[61986]: value = "task-1159948" [ 915.524061] env[61986]: _type = "Task" [ 915.524061] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.532097] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159948, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.638775] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058255} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.639571] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.640144] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a6061c-8722-49db-9f28-554d32d4d01b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.664451] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 59cfee64-ef79-4b8f-a703-c8812551d12f/59cfee64-ef79-4b8f-a703-c8812551d12f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.664789] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-757fdecc-8ea5-4027-ba0f-af989c6c4f0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.685031] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 915.685031] env[61986]: value = "task-1159949" [ 915.685031] env[61986]: _type = "Task" [ 915.685031] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.693394] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.723927] env[61986]: DEBUG nova.network.neutron [-] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.751719] env[61986]: INFO nova.compute.manager [-] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Took 1.53 seconds to deallocate network for instance. [ 915.794349] env[61986]: DEBUG nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.797724] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 915.802478] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fe848a9a-9d09-4bbf-8a9b-2dd16d871de3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.805898] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ae85f26-4163-4b48-95be-881af7d2a8dd tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.873s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.813561] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 915.813561] env[61986]: value = "task-1159950" [ 915.813561] env[61986]: _type = "Task" [ 915.813561] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.830303] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159950, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.048655] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159948, 'name': CreateVM_Task, 'duration_secs': 0.363347} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.048875] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 916.049876] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.050141] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.051277] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.051277] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f49244c-da30-4f15-b726-273ec078b951 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.057852] env[61986]: DEBUG nova.network.neutron [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.060898] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 916.060898] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527fe2a3-5f3d-81a4-ede8-707375d66d4d" [ 916.060898] env[61986]: _type = "Task" [ 916.060898] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.074814] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527fe2a3-5f3d-81a4-ede8-707375d66d4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.195236] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159949, 'name': ReconfigVM_Task, 'duration_secs': 0.269094} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.197917] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 59cfee64-ef79-4b8f-a703-c8812551d12f/59cfee64-ef79-4b8f-a703-c8812551d12f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.198776] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8518fa3-7256-482c-a042-62f82cbf8bdc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.209116] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 916.209116] env[61986]: value = "task-1159951" [ 916.209116] env[61986]: _type = "Task" [ 916.209116] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.221347] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159951, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.227986] env[61986]: INFO nova.compute.manager [-] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Took 1.43 seconds to deallocate network for instance. [ 916.251980] env[61986]: DEBUG nova.network.neutron [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Updated VIF entry in instance network info cache for port 495a4f17-b010-4427-bf0f-e72582cd1dc1. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 916.252160] env[61986]: DEBUG nova.network.neutron [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Updating instance_info_cache with network_info: [{"id": "495a4f17-b010-4427-bf0f-e72582cd1dc1", "address": "fa:16:3e:f6:2a:40", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap495a4f17-b0", "ovs_interfaceid": "495a4f17-b010-4427-bf0f-e72582cd1dc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.262943] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.328133] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159950, 'name': CreateSnapshot_Task, 'duration_secs': 0.503574} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.328133] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 916.328813] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a47db5-8868-4555-a3ae-fcd5b4398d8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.333011] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463c060f-4fa3-4967-a2bb-6acf1b4cacc6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.354323] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab734cc2-8ebf-44ba-a7e0-e5f16bdcb402 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.390121] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477ed47d-5b7a-45e8-b6cd-8f3c36618e92 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.399943] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313009e3-b3f0-460c-82e0-11b4f23dfd0e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.404918] env[61986]: DEBUG nova.network.neutron [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Updating instance_info_cache with network_info: [{"id": "9afcb8ca-5b37-41f4-a4de-a97cf229512b", "address": "fa:16:3e:60:1a:97", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9afcb8ca-5b", "ovs_interfaceid": "9afcb8ca-5b37-41f4-a4de-a97cf229512b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.418693] env[61986]: DEBUG nova.compute.provider_tree [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.538557] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.538557] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.538777] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.538962] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.539197] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.542205] env[61986]: INFO nova.compute.manager [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Terminating instance [ 916.544169] env[61986]: DEBUG nova.compute.manager [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.544378] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.545228] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf72f1e1-6693-4ace-8790-09770b4b738e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.552857] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 916.553086] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d725201-5965-4eb5-85db-1a1663d3eedf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.559120] env[61986]: DEBUG oslo_vmware.api [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 916.559120] env[61986]: value = "task-1159952" [ 916.559120] env[61986]: _type = "Task" [ 916.559120] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.572098] env[61986]: DEBUG oslo_vmware.api [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159952, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.578067] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527fe2a3-5f3d-81a4-ede8-707375d66d4d, 'name': SearchDatastore_Task, 'duration_secs': 0.015797} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.578386] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.578619] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.579652] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.579652] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.579652] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.579652] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2896e02f-0821-4614-92d5-db8af3eb01e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.589174] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.589418] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 916.590114] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0abc6659-ea79-45cb-88b6-1ce24ad97531 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.595090] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 916.595090] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5298025f-bf1a-063b-2f90-ddc4ac16a936" [ 916.595090] env[61986]: _type = "Task" [ 916.595090] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.602848] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5298025f-bf1a-063b-2f90-ddc4ac16a936, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.650462] env[61986]: DEBUG nova.compute.manager [req-2b36c67c-61d4-4aa4-88e5-f35947bcf422 req-2b3b9499-4792-4803-b6cb-ede1b217495a service nova] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Received event network-vif-deleted-18e4cc01-51d6-4a7f-b279-4427cedb88ca {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.650688] env[61986]: DEBUG nova.compute.manager [req-2b36c67c-61d4-4aa4-88e5-f35947bcf422 req-2b3b9499-4792-4803-b6cb-ede1b217495a service nova] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Received event network-vif-deleted-2deaa3c0-1890-4e85-83f6-1ca9a7007616 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.715800] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.716250] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.722489] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159951, 'name': Rename_Task, 'duration_secs': 0.154889} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.722489] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 916.722489] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84e7f90c-ea0e-4c65-aea5-9de63d40b7a6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.729773] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 916.729773] env[61986]: value = "task-1159953" [ 916.729773] env[61986]: _type = "Task" [ 916.729773] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.742015] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.742015] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159953, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.755340] env[61986]: DEBUG oslo_concurrency.lockutils [req-846816ae-db56-4cbb-895d-ac50fbd57ea0 req-71dfc976-b042-4ade-b5e9-56686e36e0f1 service nova] Releasing lock "refresh_cache-49bcac13-6dde-4a28-9131-f9a1e1d2e386" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.806114] env[61986]: DEBUG nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.841040] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.841193] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.841798] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.841798] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.841798] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.842038] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.842701] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.842701] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.842701] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.844029] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.844029] env[61986]: DEBUG nova.virt.hardware [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.844658] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e31ec21-fc61-4f93-aa22-82ba8f43c68d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.853480] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccf8868-9a8b-4109-90cb-85dd12eab77d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.863985] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 916.864619] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dec01643-2421-4cd3-b412-eb2736d7c138 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.877911] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.883699] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Creating folder: Project (baaa79c32a334572875958a3edea5e8a). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 916.885331] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24ecf7a0-436a-4f81-9858-4c6b5ad4b1f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.887092] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 916.887092] env[61986]: value = "task-1159954" [ 916.887092] env[61986]: _type = "Task" [ 916.887092] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.895579] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159954, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.896805] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Created folder: Project (baaa79c32a334572875958a3edea5e8a) in parent group-v252271. [ 916.897029] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Creating folder: Instances. Parent ref: group-v252389. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 916.897283] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d60db1a-3402-42c4-88d3-1c6fc4344eda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.907432] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Releasing lock "refresh_cache-6488f68e-7b84-4462-aef5-25d02db504f6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.907723] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Instance network_info: |[{"id": "9afcb8ca-5b37-41f4-a4de-a97cf229512b", "address": "fa:16:3e:60:1a:97", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9afcb8ca-5b", "ovs_interfaceid": "9afcb8ca-5b37-41f4-a4de-a97cf229512b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 916.908032] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Created folder: Instances in parent group-v252389. [ 916.908284] env[61986]: DEBUG oslo.service.loopingcall [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.908626] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:1a:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9afcb8ca-5b37-41f4-a4de-a97cf229512b', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.916342] env[61986]: DEBUG oslo.service.loopingcall [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.916563] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 916.916838] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 916.917068] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee8f6ef9-f059-4657-a31d-d6fa72a09954 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.931737] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95bb792f-f387-48a2-9abd-1435a768b612 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.949718] env[61986]: DEBUG nova.scheduler.client.report [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.960104] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.960104] env[61986]: value = "task-1159957" [ 916.960104] env[61986]: _type = "Task" [ 916.960104] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.960593] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.960593] env[61986]: value = "task-1159958" [ 916.960593] env[61986]: _type = "Task" [ 916.960593] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.974658] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159957, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.978016] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159958, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.068992] env[61986]: DEBUG oslo_vmware.api [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159952, 'name': PowerOffVM_Task, 'duration_secs': 0.300581} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.069330] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 917.069504] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 917.069762] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02b3c0d6-fec9-4d23-8cc2-d933f531efe9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.105749] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5298025f-bf1a-063b-2f90-ddc4ac16a936, 'name': SearchDatastore_Task, 'duration_secs': 0.00982} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.106674] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a21a6aa2-cf35-497e-876b-82b1e5637836 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.112179] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 917.112179] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274ccea-e2d7-7790-555d-93b62f2aaec3" [ 917.112179] env[61986]: _type = "Task" [ 917.112179] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.122609] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274ccea-e2d7-7790-555d-93b62f2aaec3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.128488] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 917.128831] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 917.129143] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Deleting the datastore file [datastore1] 6e0697d4-d862-497c-9bdd-dd1ef2d4272b {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.129471] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf45fe9f-f2e2-47a3-b503-10bc1c4393c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.136411] env[61986]: DEBUG oslo_vmware.api [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for the task: (returnval){ [ 917.136411] env[61986]: value = "task-1159960" [ 917.136411] env[61986]: _type = "Task" [ 917.136411] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.144624] env[61986]: DEBUG oslo_vmware.api [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.224022] env[61986]: DEBUG nova.compute.utils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.240933] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159953, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.356644] env[61986]: DEBUG nova.compute.manager [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Received event network-changed-9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.356797] env[61986]: DEBUG nova.compute.manager [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Refreshing instance network info cache due to event network-changed-9afcb8ca-5b37-41f4-a4de-a97cf229512b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 917.356975] env[61986]: DEBUG oslo_concurrency.lockutils [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] Acquiring lock "refresh_cache-6488f68e-7b84-4462-aef5-25d02db504f6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.357195] env[61986]: DEBUG oslo_concurrency.lockutils [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] Acquired lock "refresh_cache-6488f68e-7b84-4462-aef5-25d02db504f6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.357365] env[61986]: DEBUG nova.network.neutron [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Refreshing network info cache for port 9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.403547] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159954, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.456788] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.674s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.458322] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.462250] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.068s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.463358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.465563] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.802s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.467216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.469812] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.349s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.472666] env[61986]: INFO nova.compute.claims [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.492743] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159957, 'name': CreateVM_Task, 'duration_secs': 0.495} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.496621] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 917.496879] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159958, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.497342] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.497752] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.497880] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.498188] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-497216f7-ee4c-4ad4-a0cd-9c78ac3b8b5e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.503895] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 917.503895] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52195ac2-35d3-5eea-dd54-509dfb29eab2" [ 917.503895] env[61986]: _type = "Task" [ 917.503895] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.509192] env[61986]: INFO nova.scheduler.client.report [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted allocations for instance 9b539a94-6c39-46f8-b194-27047245d1f2 [ 917.510840] env[61986]: INFO nova.scheduler.client.report [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Deleted allocations for instance d72cd38b-ef14-467b-bf53-97d9e66534c8 [ 917.524302] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52195ac2-35d3-5eea-dd54-509dfb29eab2, 'name': SearchDatastore_Task, 'duration_secs': 0.011775} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.524918] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.525414] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.525510] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.525687] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.525913] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.526273] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67ef22e5-258f-4270-ac72-307682b88390 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.534787] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.534977] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 917.535918] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-816ceb98-1a56-4205-a4b8-9704d2fbab15 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.542500] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 917.542500] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52334a0b-39ce-f698-a2a3-72ea8c713d70" [ 917.542500] env[61986]: _type = "Task" [ 917.542500] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.550942] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52334a0b-39ce-f698-a2a3-72ea8c713d70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.624132] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274ccea-e2d7-7790-555d-93b62f2aaec3, 'name': SearchDatastore_Task, 'duration_secs': 0.008509} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.624416] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.624681] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 49bcac13-6dde-4a28-9131-f9a1e1d2e386/49bcac13-6dde-4a28-9131-f9a1e1d2e386.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 917.625344] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60e517eb-7b27-41df-bf4a-10faec5fa6e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.632121] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 917.632121] env[61986]: value = "task-1159961" [ 917.632121] env[61986]: _type = "Task" [ 917.632121] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.644812] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.650279] env[61986]: DEBUG oslo_vmware.api [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Task: {'id': task-1159960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124977} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.650570] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.650866] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 917.650907] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 917.651088] env[61986]: INFO nova.compute.manager [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 917.651334] env[61986]: DEBUG oslo.service.loopingcall [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.651539] env[61986]: DEBUG nova.compute.manager [-] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.651637] env[61986]: DEBUG nova.network.neutron [-] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 917.727628] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.743021] env[61986]: DEBUG oslo_vmware.api [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159953, 'name': PowerOnVM_Task, 'duration_secs': 0.919657} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.743327] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 917.743563] env[61986]: INFO nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Took 8.46 seconds to spawn the instance on the hypervisor. [ 917.743754] env[61986]: DEBUG nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.744660] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c86507d-15f0-4445-86a1-a5089c6e5677 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.799267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.799506] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.902173] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159954, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.982505] env[61986]: DEBUG nova.compute.utils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.986079] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.986267] env[61986]: DEBUG nova.network.neutron [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.011299] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159958, 'name': CreateVM_Task, 'duration_secs': 0.718706} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.012559] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 918.013457] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.013900] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.014037] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.014609] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-526f47b7-8147-4f92-8779-d9962ecc3b52 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.021033] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebd8cc2-e931-428f-8726-d58ef0987226 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "9b539a94-6c39-46f8-b194-27047245d1f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.746s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.023949] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 918.023949] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5285c456-514d-e3e6-043e-ea96e2f56209" [ 918.023949] env[61986]: _type = "Task" [ 918.023949] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.031603] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f31491e4-eb9d-41ec-9af9-3b840639fe69 tempest-MultipleCreateTestJSON-183406664 tempest-MultipleCreateTestJSON-183406664-project-member] Lock "d72cd38b-ef14-467b-bf53-97d9e66534c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.503s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.036184] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5285c456-514d-e3e6-043e-ea96e2f56209, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.060323] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52334a0b-39ce-f698-a2a3-72ea8c713d70, 'name': SearchDatastore_Task, 'duration_secs': 0.009103} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.062825] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f486a46a-3518-4d48-9043-53a612bfd3fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.072013] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 918.072013] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c0753-5088-056a-1bb6-3b44f7eb1b92" [ 918.072013] env[61986]: _type = "Task" [ 918.072013] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.085561] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c0753-5088-056a-1bb6-3b44f7eb1b92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.092788] env[61986]: DEBUG nova.policy [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c637e468ca794a1bac4da331337a1793', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0250d1bd7f4147a19dcb32797b5d794b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 918.145544] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159961, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.216269] env[61986]: DEBUG nova.network.neutron [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Updated VIF entry in instance network info cache for port 9afcb8ca-5b37-41f4-a4de-a97cf229512b. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.216891] env[61986]: DEBUG nova.network.neutron [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Updating instance_info_cache with network_info: [{"id": "9afcb8ca-5b37-41f4-a4de-a97cf229512b", "address": "fa:16:3e:60:1a:97", "network": {"id": "29de0f53-748d-4364-a05d-e240135ec071", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "f93eeac37dc4477cb79aae49b3cd744b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9afcb8ca-5b", "ovs_interfaceid": "9afcb8ca-5b37-41f4-a4de-a97cf229512b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.271422] env[61986]: INFO nova.compute.manager [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Took 37.96 seconds to build instance. [ 918.401478] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1159954, 'name': CloneVM_Task, 'duration_secs': 1.276558} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.401783] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Created linked-clone VM from snapshot [ 918.403259] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5d9e9b-ec9c-4382-a094-0134c8524b91 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.411685] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Uploading image b36a707b-06ac-40f7-8b1a-e866a1c488a2 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 918.446130] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 918.446130] env[61986]: value = "vm-252391" [ 918.446130] env[61986]: _type = "VirtualMachine" [ 918.446130] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 918.446448] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b7b90560-1c16-40f6-8e8a-17f82be50e3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.454114] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease: (returnval){ [ 918.454114] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529b077b-0d83-78a5-45da-5cf09e1826b8" [ 918.454114] env[61986]: _type = "HttpNfcLease" [ 918.454114] env[61986]: } obtained for exporting VM: (result){ [ 918.454114] env[61986]: value = "vm-252391" [ 918.454114] env[61986]: _type = "VirtualMachine" [ 918.454114] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 918.454497] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the lease: (returnval){ [ 918.454497] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529b077b-0d83-78a5-45da-5cf09e1826b8" [ 918.454497] env[61986]: _type = "HttpNfcLease" [ 918.454497] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 918.460875] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.460875] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529b077b-0d83-78a5-45da-5cf09e1826b8" [ 918.460875] env[61986]: _type = "HttpNfcLease" [ 918.460875] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 918.461680] env[61986]: DEBUG nova.network.neutron [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Successfully created port: c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.499083] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.537168] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5285c456-514d-e3e6-043e-ea96e2f56209, 'name': SearchDatastore_Task, 'duration_secs': 0.069527} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.537168] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.537168] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.537168] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.537168] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.537168] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.537168] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6cda860-3b4a-4e1f-955c-e8ed50581eb9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.542961] env[61986]: DEBUG nova.network.neutron [-] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.547542] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.547981] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 918.551769] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63d8173f-3eb2-442b-832f-2a7e5be9a5ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.560963] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 918.560963] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cf993b-f24d-ebb5-0e68-26c6690d4bde" [ 918.560963] env[61986]: _type = "Task" [ 918.560963] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.568861] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cf993b-f24d-ebb5-0e68-26c6690d4bde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.583213] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c0753-5088-056a-1bb6-3b44f7eb1b92, 'name': SearchDatastore_Task, 'duration_secs': 0.034578} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.583634] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.584024] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 918.584419] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfda8f3a-6d9f-44a2-8acd-89139a6c822f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.594467] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 918.594467] env[61986]: value = "task-1159963" [ 918.594467] env[61986]: _type = "Task" [ 918.594467] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.605130] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.644764] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159961, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721018} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.647710] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 49bcac13-6dde-4a28-9131-f9a1e1d2e386/49bcac13-6dde-4a28-9131-f9a1e1d2e386.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 918.647937] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.648574] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-279d7e17-2d50-46cf-a69d-012e57f9b351 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.654474] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 918.654474] env[61986]: value = "task-1159964" [ 918.654474] env[61986]: _type = "Task" [ 918.654474] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.668848] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.720754] env[61986]: DEBUG oslo_concurrency.lockutils [req-d2f16615-9122-4676-bd9a-d19524982628 req-82da58ab-4e4d-45fb-830a-2aa26892ac3d service nova] Releasing lock "refresh_cache-6488f68e-7b84-4462-aef5-25d02db504f6" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.775653] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7e8ef4ad-517b-4642-9001-b10815d4052c tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.221s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.825731] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.826047] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.826309] env[61986]: INFO nova.compute.manager [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Attaching volume 47fff893-4caa-4a2a-ba29-4ee2766f6406 to /dev/sdc [ 918.884764] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95833930-a56c-4432-b56d-82870f584c77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.895424] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe54a73b-8f77-4f56-b22f-2920208aea02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.913079] env[61986]: DEBUG nova.virt.block_device [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating existing volume attachment record: 797d3236-e1a6-496b-8c6c-2fff41813ce7 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 918.969321] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.969321] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529b077b-0d83-78a5-45da-5cf09e1826b8" [ 918.969321] env[61986]: _type = "HttpNfcLease" [ 918.969321] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 918.969579] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 918.969579] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529b077b-0d83-78a5-45da-5cf09e1826b8" [ 918.969579] env[61986]: _type = "HttpNfcLease" [ 918.969579] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 918.970388] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe61c3e-e5b0-459b-935d-ecda2f8b6e4a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.981428] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525a6f18-5ebc-39f7-a5e1-b0b5cc9d89e3/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 918.981672] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525a6f18-5ebc-39f7-a5e1-b0b5cc9d89e3/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 919.048322] env[61986]: INFO nova.compute.manager [-] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Took 1.40 seconds to deallocate network for instance. [ 919.071857] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cf993b-f24d-ebb5-0e68-26c6690d4bde, 'name': SearchDatastore_Task, 'duration_secs': 0.052823} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.073246] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aca0d112-c7f4-42e2-842d-5d897c7a4b82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.079806] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 919.079806] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fde683-d91d-df71-4af2-d539117148a9" [ 919.079806] env[61986]: _type = "Task" [ 919.079806] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.091818] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fde683-d91d-df71-4af2-d539117148a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.106389] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.131280] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ac48af-9e1d-4bbf-a6cb-11467b5234e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.139457] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48b714b-f377-4490-99a5-8a5ba998c3a5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.178269] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd7b0ea-0754-475d-8285-e5cf6755d337 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.181693] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-11082e46-12fb-40ec-8a38-1b76a81bd750 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.193783] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4e904c-17bd-4ccb-98cf-6c0fab429b48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.198143] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.293112} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.200139] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.201779] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490a5e4b-2adc-4e6a-8540-065a76c69ca6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.212791] env[61986]: DEBUG nova.compute.provider_tree [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.236761] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 49bcac13-6dde-4a28-9131-f9a1e1d2e386/49bcac13-6dde-4a28-9131-f9a1e1d2e386.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.242490] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3535bd3c-b3dc-4bd3-a6b6-68e03513457e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.266046] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 919.266046] env[61986]: value = "task-1159966" [ 919.266046] env[61986]: _type = "Task" [ 919.266046] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.270920] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159966, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.279031] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.414698] env[61986]: DEBUG nova.compute.manager [req-62a24a4d-96e2-409c-8631-92d19c1620e2 req-45fa4d24-c4ec-4735-9a6f-5d014b3732c7 service nova] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Received event network-vif-deleted-a9fa57e7-ab8e-4c91-b1b1-20aeba0808f3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.550508] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.561512] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.582972] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.583375] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.583684] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.583992] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.584263] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.584455] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.584751] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.584940] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.585137] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.585349] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.586434] env[61986]: DEBUG nova.virt.hardware [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.587530] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a99221-1c1d-41fc-8fca-fb56422beeaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.607091] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fde683-d91d-df71-4af2-d539117148a9, 'name': SearchDatastore_Task, 'duration_secs': 0.025232} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.608933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d821fcaf-3540-4301-8edd-4d266fcbf8b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.613448] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.613821] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6488f68e-7b84-4462-aef5-25d02db504f6/6488f68e-7b84-4462-aef5-25d02db504f6.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 919.617865] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb9ae520-322e-4b52-80e7-64d307e5d8e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.620136] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.995589} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.621406] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 919.621406] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.621881] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f6e20ac-a19d-44b7-8a8e-63adfc115c63 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.633402] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 919.633402] env[61986]: value = "task-1159967" [ 919.633402] env[61986]: _type = "Task" [ 919.633402] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.638658] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 919.638658] env[61986]: value = "task-1159968" [ 919.638658] env[61986]: _type = "Task" [ 919.638658] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.645747] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.651384] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.719436] env[61986]: DEBUG nova.scheduler.client.report [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.775301] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159966, 'name': ReconfigVM_Task, 'duration_secs': 0.346411} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.776119] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 49bcac13-6dde-4a28-9131-f9a1e1d2e386/49bcac13-6dde-4a28-9131-f9a1e1d2e386.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.777464] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a0e170a-e771-47cc-8216-822544b35a4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.792127] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 919.792127] env[61986]: value = "task-1159969" [ 919.792127] env[61986]: _type = "Task" [ 919.792127] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.805957] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159969, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.818209] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.115377] env[61986]: DEBUG nova.network.neutron [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Successfully updated port: c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.147125] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499532} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.147889] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6488f68e-7b84-4462-aef5-25d02db504f6/6488f68e-7b84-4462-aef5-25d02db504f6.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 920.148208] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.148828] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07531140-978f-47b4-ae8f-df7bb6575c75 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.154156] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069163} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.154796] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.155629] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ed007c-9bca-4db0-bc21-16f74a78f13a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.160235] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 920.160235] env[61986]: value = "task-1159970" [ 920.160235] env[61986]: _type = "Task" [ 920.160235] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.180358] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.181314] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbca3f39-1f0f-4aa4-ae39-c987d9eef74b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.201127] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.208645] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 920.208645] env[61986]: value = "task-1159971" [ 920.208645] env[61986]: _type = "Task" [ 920.208645] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.217430] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.222710] env[61986]: DEBUG nova.compute.manager [req-5160109c-27fd-4e9d-9827-50ba973da831 req-5b175e49-7e3d-44a1-8dc8-e8454c31f45a service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Received event network-vif-plugged-c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.225042] env[61986]: DEBUG oslo_concurrency.lockutils [req-5160109c-27fd-4e9d-9827-50ba973da831 req-5b175e49-7e3d-44a1-8dc8-e8454c31f45a service nova] Acquiring lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.225042] env[61986]: DEBUG oslo_concurrency.lockutils [req-5160109c-27fd-4e9d-9827-50ba973da831 req-5b175e49-7e3d-44a1-8dc8-e8454c31f45a service nova] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.225042] env[61986]: DEBUG oslo_concurrency.lockutils [req-5160109c-27fd-4e9d-9827-50ba973da831 req-5b175e49-7e3d-44a1-8dc8-e8454c31f45a service nova] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.225042] env[61986]: DEBUG nova.compute.manager [req-5160109c-27fd-4e9d-9827-50ba973da831 req-5b175e49-7e3d-44a1-8dc8-e8454c31f45a service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] No waiting events found dispatching network-vif-plugged-c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.225042] env[61986]: WARNING nova.compute.manager [req-5160109c-27fd-4e9d-9827-50ba973da831 req-5b175e49-7e3d-44a1-8dc8-e8454c31f45a service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Received unexpected event network-vif-plugged-c4411312-86a5-48dd-a108-5c5bc29d5f3e for instance with vm_state building and task_state spawning. [ 920.225902] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.756s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.227074] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.229086] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.506s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.229293] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.229495] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 920.229910] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.013s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.232074] env[61986]: INFO nova.compute.claims [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.237321] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c432377d-c04e-438e-9086-a4b633e0af8f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.247593] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93c54aa-6828-483f-b56f-95be9c885d47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.267518] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f5e4b4-2a4c-4909-8593-509c0abdf257 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.274632] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c210b5df-06fe-4114-bc44-fbea9c869190 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.313218] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179676MB free_disk=162GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 920.313522] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.322817] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159969, 'name': Rename_Task, 'duration_secs': 0.159591} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.323194] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 920.323532] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2936168-03e9-430d-900b-dfadb5d0d287 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.330700] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 920.330700] env[61986]: value = "task-1159972" [ 920.330700] env[61986]: _type = "Task" [ 920.330700] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.339096] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.621131] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.621333] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.621540] env[61986]: DEBUG nova.network.neutron [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.670180] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062803} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.670478] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.671352] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73795b70-b183-416f-b87c-26da8f791b23 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.695595] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 6488f68e-7b84-4462-aef5-25d02db504f6/6488f68e-7b84-4462-aef5-25d02db504f6.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.695888] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ea53d6d-21c4-4039-8b29-fa2b266fabd1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.725806] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159971, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.730302] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 920.730302] env[61986]: value = "task-1159973" [ 920.730302] env[61986]: _type = "Task" [ 920.730302] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.743198] env[61986]: DEBUG nova.compute.utils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.744501] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159973, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.745111] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.745622] env[61986]: DEBUG nova.network.neutron [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 920.842339] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159972, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.843688] env[61986]: DEBUG nova.policy [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4675c2084b6463bb47c93865f5cec46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f746f57f466f480c958d8b15bd2ce8fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 921.172562] env[61986]: DEBUG nova.network.neutron [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.191633] env[61986]: DEBUG nova.network.neutron [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Successfully created port: ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.228759] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159971, 'name': ReconfigVM_Task, 'duration_secs': 0.641229} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.229208] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Reconfigured VM instance instance-0000004b to attach disk [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.229863] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e49deb7-9a5a-4963-a1f1-cbe9641e3448 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.242634] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159973, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.246601] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 921.246601] env[61986]: value = "task-1159975" [ 921.246601] env[61986]: _type = "Task" [ 921.246601] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.249909] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.265633] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159975, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.349976] env[61986]: DEBUG oslo_vmware.api [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159972, 'name': PowerOnVM_Task, 'duration_secs': 0.621172} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.350623] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 921.350717] env[61986]: INFO nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Took 9.42 seconds to spawn the instance on the hypervisor. [ 921.350967] env[61986]: DEBUG nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.351925] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3c1958-156a-4df5-8a66-7a9abd551ae5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.486605] env[61986]: DEBUG nova.network.neutron [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updating instance_info_cache with network_info: [{"id": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "address": "fa:16:3e:41:aa:c3", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4411312-86", "ovs_interfaceid": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.749242] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159973, 'name': ReconfigVM_Task, 'duration_secs': 0.637335} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.754368] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 6488f68e-7b84-4462-aef5-25d02db504f6/6488f68e-7b84-4462-aef5-25d02db504f6.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.759482] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41bda408-39f2-4605-beb7-c8dd0ebd2aa2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.772476] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159975, 'name': Rename_Task, 'duration_secs': 0.170044} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.773973] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 921.774338] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 921.774338] env[61986]: value = "task-1159976" [ 921.774338] env[61986]: _type = "Task" [ 921.774338] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.776989] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aad09e9e-45d7-4085-b1fb-baefd72052b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.787982] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159976, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.789584] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 921.789584] env[61986]: value = "task-1159977" [ 921.789584] env[61986]: _type = "Task" [ 921.789584] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.800340] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159977, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.833692] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68df6d93-d322-4039-88bd-7edac65d9386 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.843562] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea9aed9-28f1-4954-95d3-f1081a53df7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.881244] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5851e0f2-2465-4518-bfa1-e5dc09b3aacc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.883848] env[61986]: INFO nova.compute.manager [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Took 38.20 seconds to build instance. [ 921.891237] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6d1235-26fb-4792-901f-7f0f4bac77c8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.906395] env[61986]: DEBUG nova.compute.provider_tree [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.971453] env[61986]: INFO nova.compute.manager [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Rebuilding instance [ 921.990665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.993078] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Instance network_info: |[{"id": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "address": "fa:16:3e:41:aa:c3", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4411312-86", "ovs_interfaceid": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.993078] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:aa:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4411312-86a5-48dd-a108-5c5bc29d5f3e', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.004427] env[61986]: DEBUG oslo.service.loopingcall [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.007115] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.012391] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8dd5951-4e60-4d88-904d-dedfb42ef82a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.048826] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.048826] env[61986]: value = "task-1159978" [ 922.048826] env[61986]: _type = "Task" [ 922.048826] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.058215] env[61986]: DEBUG nova.compute.manager [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.059265] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08087492-3b78-4c4e-9b80-a8a9f253a614 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.065376] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159978, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.266890] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.288532] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159976, 'name': Rename_Task, 'duration_secs': 0.214982} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.290797] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.291046] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.291214] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.291405] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.291765] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.291765] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.291896] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.292066] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.292241] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.292405] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.292649] env[61986]: DEBUG nova.virt.hardware [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.292967] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 922.293788] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829ad056-a1bd-437c-a9a6-77c8b5a56ab7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.296339] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18baaedf-b90c-4dd2-9a10-ef95dae40ba6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.310062] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76335ed3-eb71-4ba4-81b8-3527ef3771f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.314245] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 922.314245] env[61986]: value = "task-1159979" [ 922.314245] env[61986]: _type = "Task" [ 922.314245] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.314478] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159977, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.331347] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159979, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.385975] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1481c4db-41ce-4907-bfda-ecf2414228f2 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.865s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.410577] env[61986]: DEBUG nova.scheduler.client.report [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.420315] env[61986]: DEBUG nova.compute.manager [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Received event network-changed-c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.420587] env[61986]: DEBUG nova.compute.manager [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Refreshing instance network info cache due to event network-changed-c4411312-86a5-48dd-a108-5c5bc29d5f3e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.420848] env[61986]: DEBUG oslo_concurrency.lockutils [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] Acquiring lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.421063] env[61986]: DEBUG oslo_concurrency.lockutils [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] Acquired lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.421252] env[61986]: DEBUG nova.network.neutron [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Refreshing network info cache for port c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.561749] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159978, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.574514] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.574951] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-163fa4ba-dd9e-4190-b5e5-4f18d608da6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.582915] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 922.582915] env[61986]: value = "task-1159980" [ 922.582915] env[61986]: _type = "Task" [ 922.582915] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.592162] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.805983] env[61986]: DEBUG oslo_vmware.api [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159977, 'name': PowerOnVM_Task, 'duration_secs': 0.673094} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.806330] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.806547] env[61986]: INFO nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Took 6.00 seconds to spawn the instance on the hypervisor. [ 922.806738] env[61986]: DEBUG nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.807791] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625d6297-24f2-4f73-a006-d0e1deaf26a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.832994] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159979, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.917214] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.917214] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 922.920152] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.415s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.925417] env[61986]: INFO nova.compute.claims [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.067337] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159978, 'name': CreateVM_Task, 'duration_secs': 0.644658} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.069157] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.069157] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.069157] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.069392] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.070213] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a307c291-ce21-426f-9f0c-a68ca3b76509 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.076374] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 923.076374] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529dc811-0f70-3357-fc3f-3a311f1389a9" [ 923.076374] env[61986]: _type = "Task" [ 923.076374] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.088490] env[61986]: DEBUG nova.network.neutron [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Successfully updated port: ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.102186] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529dc811-0f70-3357-fc3f-3a311f1389a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.106989] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159980, 'name': PowerOffVM_Task, 'duration_secs': 0.238324} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.107470] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 923.108316] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 923.108914] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b2e9ee6-891a-4777-b428-6499b3a24b81 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.116311] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 923.116311] env[61986]: value = "task-1159981" [ 923.116311] env[61986]: _type = "Task" [ 923.116311] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.128724] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 923.129133] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 923.130368] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252302', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'name': 'volume-6d6d6476-0f93-4404-b58d-3092cea80b1b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98337bb6-9502-4d4c-af00-028659b246bf', 'attached_at': '', 'detached_at': '', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'serial': '6d6d6476-0f93-4404-b58d-3092cea80b1b'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 923.130368] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6d5c2b-2cac-41f3-94b5-ce86e1a0af05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.152985] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469a1eea-0d37-41e2-968d-218eb4fa6a7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.161113] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b57f5c-2192-4a9b-9988-899400f3005b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.180412] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937bcbde-e149-40d4-92c4-969159c20bac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.199112] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] The volume has not been displaced from its original location: [datastore2] volume-6d6d6476-0f93-4404-b58d-3092cea80b1b/volume-6d6d6476-0f93-4404-b58d-3092cea80b1b.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 923.205237] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Reconfiguring VM instance instance-00000040 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 923.209224] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9263552d-79eb-41ea-a7b7-29f777c4c4b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.227387] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 923.227387] env[61986]: value = "task-1159982" [ 923.227387] env[61986]: _type = "Task" [ 923.227387] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.237065] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159982, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.257460] env[61986]: DEBUG nova.network.neutron [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updated VIF entry in instance network info cache for port c4411312-86a5-48dd-a108-5c5bc29d5f3e. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.257950] env[61986]: DEBUG nova.network.neutron [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updating instance_info_cache with network_info: [{"id": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "address": "fa:16:3e:41:aa:c3", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4411312-86", "ovs_interfaceid": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.332500] env[61986]: DEBUG oslo_vmware.api [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1159979, 'name': PowerOnVM_Task, 'duration_secs': 0.721259} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.334006] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 923.334248] env[61986]: INFO nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Took 9.20 seconds to spawn the instance on the hypervisor. [ 923.334430] env[61986]: DEBUG nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.335190] env[61986]: INFO nova.compute.manager [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Took 33.50 seconds to build instance. [ 923.336709] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e76d0cd-340a-4cda-b168-724ed5939876 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.430508] env[61986]: DEBUG nova.compute.utils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.432877] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 923.432877] env[61986]: DEBUG nova.network.neutron [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 923.479837] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 923.480210] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252394', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'name': 'volume-47fff893-4caa-4a2a-ba29-4ee2766f6406', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'serial': '47fff893-4caa-4a2a-ba29-4ee2766f6406'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 923.481537] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7f95f7-8f2f-409f-a766-02408392fd4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.500962] env[61986]: DEBUG nova.policy [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f45c3c8e766d4483940d34edde7ae965', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fe9c887c0d3485f8e6f01785b2a4743', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 923.505812] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee042ee4-bd02-4de8-8dee-35dc4e85801a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.536522] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] volume-47fff893-4caa-4a2a-ba29-4ee2766f6406/volume-47fff893-4caa-4a2a-ba29-4ee2766f6406.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.537585] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be3de313-cb48-40f0-854d-4c6e6cecf7ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.560327] env[61986]: DEBUG oslo_vmware.api [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 923.560327] env[61986]: value = "task-1159983" [ 923.560327] env[61986]: _type = "Task" [ 923.560327] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.575237] env[61986]: DEBUG oslo_vmware.api [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159983, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.589183] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529dc811-0f70-3357-fc3f-3a311f1389a9, 'name': SearchDatastore_Task, 'duration_secs': 0.024038} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.589530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.589790] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.590077] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.590235] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.590445] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.591028] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90caa95f-dd9e-4502-ae31-9825ff9b69a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.597182] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.597323] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.597496] env[61986]: DEBUG nova.network.neutron [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.601625] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.601625] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.602476] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c90a0f0-7833-4285-add9-a7b7b65bc6b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.609226] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 923.609226] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228c360-6ae6-9900-677c-3a69e03f3872" [ 923.609226] env[61986]: _type = "Task" [ 923.609226] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.619343] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228c360-6ae6-9900-677c-3a69e03f3872, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.739088] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159982, 'name': ReconfigVM_Task, 'duration_secs': 0.24234} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.739482] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Reconfigured VM instance instance-00000040 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 923.750082] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0df9fcc2-a4ac-41d0-83a4-c0f9817ce644 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.768316] env[61986]: DEBUG oslo_concurrency.lockutils [req-4b960dd6-e7d1-4926-9a0c-314ec88746cc req-e1a97dc2-96f2-4a84-93c8-a47bb0198430 service nova] Releasing lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.775194] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 923.775194] env[61986]: value = "task-1159984" [ 923.775194] env[61986]: _type = "Task" [ 923.775194] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.784970] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159984, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.840484] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ff3ec2e6-cf7b-4dcf-8c00-f21b8789f103 tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "dea128a9-aed0-40b2-ae17-c068ea8e3452" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.780s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.844099] env[61986]: DEBUG nova.network.neutron [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Successfully created port: 8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.858733] env[61986]: INFO nova.compute.manager [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Took 40.18 seconds to build instance. [ 923.939182] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 924.011720] env[61986]: DEBUG nova.compute.manager [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.013273] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6096e8-9037-4865-b575-5f7d9f18f499 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.075161] env[61986]: DEBUG oslo_vmware.api [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159983, 'name': ReconfigVM_Task, 'duration_secs': 0.511843} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.075161] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfigured VM instance instance-0000003a to attach disk [datastore1] volume-47fff893-4caa-4a2a-ba29-4ee2766f6406/volume-47fff893-4caa-4a2a-ba29-4ee2766f6406.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.080311] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bdc4b61-f1f5-470e-9383-9f82cb26ae82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.102472] env[61986]: DEBUG oslo_vmware.api [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 924.102472] env[61986]: value = "task-1159985" [ 924.102472] env[61986]: _type = "Task" [ 924.102472] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.116656] env[61986]: DEBUG oslo_vmware.api [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.122982] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228c360-6ae6-9900-677c-3a69e03f3872, 'name': SearchDatastore_Task, 'duration_secs': 0.011726} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.127307] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ea5d50c-061a-46ec-8f34-b7d8bbc714d2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.132175] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 924.132175] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ad4646-9991-bef5-e1cd-361bb99bbdfa" [ 924.132175] env[61986]: _type = "Task" [ 924.132175] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.141964] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ad4646-9991-bef5-e1cd-361bb99bbdfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.147191] env[61986]: DEBUG nova.network.neutron [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.287292] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159984, 'name': ReconfigVM_Task, 'duration_secs': 0.16378} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.290385] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252302', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'name': 'volume-6d6d6476-0f93-4404-b58d-3092cea80b1b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98337bb6-9502-4d4c-af00-028659b246bf', 'attached_at': '', 'detached_at': '', 'volume_id': '6d6d6476-0f93-4404-b58d-3092cea80b1b', 'serial': '6d6d6476-0f93-4404-b58d-3092cea80b1b'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 924.290724] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 924.291832] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5c3e4c-179e-47c8-b98a-cc9f3a774175 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.299541] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 924.299854] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28333581-f82b-4575-b460-7d51cae26aa2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.319322] env[61986]: DEBUG nova.network.neutron [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Updating instance_info_cache with network_info: [{"id": "ffee6438-b805-4a00-94a9-83fa6e39a230", "address": "fa:16:3e:ee:2b:32", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffee6438-b8", "ovs_interfaceid": "ffee6438-b805-4a00-94a9-83fa6e39a230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.364951] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d19f4d63-04e4-48df-b408-68331757e6a8 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "6488f68e-7b84-4462-aef5-25d02db504f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.423s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.365267] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 924.365692] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 924.365797] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Deleting the datastore file [datastore2] 98337bb6-9502-4d4c-af00-028659b246bf {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.366237] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f8bdd13-6bb8-440a-92bc-a33e9dca971d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.378748] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for the task: (returnval){ [ 924.378748] env[61986]: value = "task-1159987" [ 924.378748] env[61986]: _type = "Task" [ 924.378748] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.391723] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159987, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.446663] env[61986]: INFO nova.virt.block_device [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Booting with volume 5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5 at /dev/sda [ 924.467911] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09347e2-a0e7-449f-94f9-4bd5bbe4808b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.470938] env[61986]: INFO nova.compute.manager [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Rebuilding instance [ 924.476423] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9152a8ac-498f-4826-ad03-14e1f84a40da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.511862] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fd8f17-93df-4331-9109-b4af058c84b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.515064] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa9e0e09-5b7e-4440-85a5-1efb2ac00955 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.522785] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4c23ab-6249-4fb6-b231-b79e50e2a102 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.532774] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf2a686-2469-43df-81a7-c0d2723c9706 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.546408] env[61986]: INFO nova.compute.manager [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] instance snapshotting [ 924.559918] env[61986]: DEBUG nova.compute.provider_tree [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.565016] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d64d97-a963-482e-a3ec-f7dae562fb7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.584110] env[61986]: DEBUG nova.compute.manager [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.585024] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f964eed-499a-4ffa-a291-6ef460433f73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.587439] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f02882d-5b9a-47b1-922d-1fc7ad090109 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.608430] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a8168c-aab1-413f-bd99-2cf258b9cd5a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.619490] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b188ea09-f38e-4364-8ee9-e45a36b26e98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.639509] env[61986]: DEBUG oslo_vmware.api [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159985, 'name': ReconfigVM_Task, 'duration_secs': 0.191452} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.646302] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252394', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'name': 'volume-47fff893-4caa-4a2a-ba29-4ee2766f6406', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'serial': '47fff893-4caa-4a2a-ba29-4ee2766f6406'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 924.652241] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ad4646-9991-bef5-e1cd-361bb99bbdfa, 'name': SearchDatastore_Task, 'duration_secs': 0.020761} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.656021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.656021] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/93c4f104-1812-4bb7-bfa7-cbf70a19ff51.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.656021] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8303ae71-1f62-4924-8de6-b5e4ad715eba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.672752] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c335df4-8ea1-493e-8796-2d456879f45c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.676664] env[61986]: DEBUG nova.compute.manager [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Received event network-vif-plugged-ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.676938] env[61986]: DEBUG oslo_concurrency.lockutils [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] Acquiring lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.677170] env[61986]: DEBUG oslo_concurrency.lockutils [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.677344] env[61986]: DEBUG oslo_concurrency.lockutils [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.677515] env[61986]: DEBUG nova.compute.manager [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] No waiting events found dispatching network-vif-plugged-ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.677682] env[61986]: WARNING nova.compute.manager [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Received unexpected event network-vif-plugged-ffee6438-b805-4a00-94a9-83fa6e39a230 for instance with vm_state building and task_state spawning. [ 924.677987] env[61986]: DEBUG nova.compute.manager [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Received event network-changed-ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.678172] env[61986]: DEBUG nova.compute.manager [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Refreshing instance network info cache due to event network-changed-ffee6438-b805-4a00-94a9-83fa6e39a230. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 924.678344] env[61986]: DEBUG oslo_concurrency.lockutils [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] Acquiring lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.680503] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 924.680503] env[61986]: value = "task-1159988" [ 924.680503] env[61986]: _type = "Task" [ 924.680503] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.690572] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c6b3a2-b710-437b-bd3a-17b4bcea6dea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.696775] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.709469] env[61986]: DEBUG nova.virt.block_device [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updating existing volume attachment record: 14b65da0-560d-4c32-8e2d-b0fd271f1ab4 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 924.822634] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.823134] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Instance network_info: |[{"id": "ffee6438-b805-4a00-94a9-83fa6e39a230", "address": "fa:16:3e:ee:2b:32", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffee6438-b8", "ovs_interfaceid": "ffee6438-b805-4a00-94a9-83fa6e39a230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.823356] env[61986]: DEBUG oslo_concurrency.lockutils [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] Acquired lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.823555] env[61986]: DEBUG nova.network.neutron [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Refreshing network info cache for port ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.827940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:2b:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ffee6438-b805-4a00-94a9-83fa6e39a230', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.833599] env[61986]: DEBUG oslo.service.loopingcall [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.834426] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.834680] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-542b748e-fdf9-447e-9dd1-2c2a8f0ead6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.857903] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.857903] env[61986]: value = "task-1159989" [ 924.857903] env[61986]: _type = "Task" [ 924.857903] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.868353] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159989, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.889705] env[61986]: DEBUG oslo_vmware.api [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Task: {'id': task-1159987, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082008} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.889994] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.890449] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 924.890449] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 924.950456] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 924.950456] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1cafa1a9-3db1-4ae2-bda9-234228dc5c0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.959924] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f33447-9686-474e-a4d2-37eac09b2eac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.997432] env[61986]: ERROR nova.compute.manager [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Failed to detach volume 6d6d6476-0f93-4404-b58d-3092cea80b1b from /dev/sda: nova.exception.InstanceNotFound: Instance 98337bb6-9502-4d4c-af00-028659b246bf could not be found. [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Traceback (most recent call last): [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self.driver.rebuild(**kwargs) [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] raise NotImplementedError() [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] NotImplementedError [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] During handling of the above exception, another exception occurred: [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Traceback (most recent call last): [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self.driver.detach_volume(context, old_connection_info, [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 552, in detach_volume [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] return self._volumeops.detach_volume(connection_info, instance) [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._detach_volume_vmdk(connection_info, instance) [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] stable_ref.fetch_moref(session) [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] raise exception.InstanceNotFound(instance_id=self._uuid) [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] nova.exception.InstanceNotFound: Instance 98337bb6-9502-4d4c-af00-028659b246bf could not be found. [ 924.997432] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 925.069468] env[61986]: DEBUG nova.scheduler.client.report [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.139612] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 925.140025] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9b4f974-4381-446d-807e-3021324dd7aa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.148281] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 925.152906] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-316a7322-d795-4dd8-a7ca-092cfb44ab15 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.155677] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 925.155677] env[61986]: value = "task-1159990" [ 925.155677] env[61986]: _type = "Task" [ 925.155677] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.166129] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 925.166129] env[61986]: value = "task-1159991" [ 925.166129] env[61986]: _type = "Task" [ 925.166129] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.168126] env[61986]: DEBUG nova.compute.utils [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Build of instance 98337bb6-9502-4d4c-af00-028659b246bf aborted: Failed to rebuild volume backed instance. {{(pid=61986) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 925.175059] env[61986]: ERROR nova.compute.manager [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 98337bb6-9502-4d4c-af00-028659b246bf aborted: Failed to rebuild volume backed instance. [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Traceback (most recent call last): [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self.driver.rebuild(**kwargs) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] raise NotImplementedError() [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] NotImplementedError [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] During handling of the above exception, another exception occurred: [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Traceback (most recent call last): [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._detach_root_volume(context, instance, root_bdm) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] with excutils.save_and_reraise_exception(): [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self.force_reraise() [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] raise self.value [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self.driver.detach_volume(context, old_connection_info, [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 552, in detach_volume [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] return self._volumeops.detach_volume(connection_info, instance) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._detach_volume_vmdk(connection_info, instance) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] stable_ref.fetch_moref(session) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] raise exception.InstanceNotFound(instance_id=self._uuid) [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] nova.exception.InstanceNotFound: Instance 98337bb6-9502-4d4c-af00-028659b246bf could not be found. [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] During handling of the above exception, another exception occurred: [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Traceback (most recent call last): [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] yield [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 925.175059] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._do_rebuild_instance_with_claim( [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._do_rebuild_instance( [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._rebuild_default_impl(**kwargs) [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] self._rebuild_volume_backed_instance( [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] raise exception.BuildAbortException( [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] nova.exception.BuildAbortException: Build of instance 98337bb6-9502-4d4c-af00-028659b246bf aborted: Failed to rebuild volume backed instance. [ 925.177070] env[61986]: ERROR nova.compute.manager [instance: 98337bb6-9502-4d4c-af00-028659b246bf] [ 925.183710] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.192179] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159991, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.197740] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472758} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.198145] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/93c4f104-1812-4bb7-bfa7-cbf70a19ff51.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.198434] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.198712] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03e53db1-b08c-4a69-92ed-8887b4b93b34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.208133] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 925.208133] env[61986]: value = "task-1159992" [ 925.208133] env[61986]: _type = "Task" [ 925.208133] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.217256] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159992, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.368923] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159989, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.552261] env[61986]: DEBUG nova.compute.manager [req-4b886e80-3832-41c1-b61e-461736c5250e req-713ef6a5-5bd8-4730-a25e-f29a9de48007 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Received event network-vif-plugged-8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.552747] env[61986]: DEBUG oslo_concurrency.lockutils [req-4b886e80-3832-41c1-b61e-461736c5250e req-713ef6a5-5bd8-4730-a25e-f29a9de48007 service nova] Acquiring lock "5170aa51-3307-42b1-b0dd-645dd4036e5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.553092] env[61986]: DEBUG oslo_concurrency.lockutils [req-4b886e80-3832-41c1-b61e-461736c5250e req-713ef6a5-5bd8-4730-a25e-f29a9de48007 service nova] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.553270] env[61986]: DEBUG oslo_concurrency.lockutils [req-4b886e80-3832-41c1-b61e-461736c5250e req-713ef6a5-5bd8-4730-a25e-f29a9de48007 service nova] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.553589] env[61986]: DEBUG nova.compute.manager [req-4b886e80-3832-41c1-b61e-461736c5250e req-713ef6a5-5bd8-4730-a25e-f29a9de48007 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] No waiting events found dispatching network-vif-plugged-8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.553820] env[61986]: WARNING nova.compute.manager [req-4b886e80-3832-41c1-b61e-461736c5250e req-713ef6a5-5bd8-4730-a25e-f29a9de48007 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Received unexpected event network-vif-plugged-8157a838-3c36-4936-a187-460e14d7f129 for instance with vm_state building and task_state block_device_mapping. [ 925.574743] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.575348] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.578127] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.887s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.578503] env[61986]: DEBUG nova.objects.instance [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lazy-loading 'resources' on Instance uuid a101dc3c-ca6b-4a72-a9b4-051b077a10fd {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.603262] env[61986]: DEBUG nova.network.neutron [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Updated VIF entry in instance network info cache for port ffee6438-b805-4a00-94a9-83fa6e39a230. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.603262] env[61986]: DEBUG nova.network.neutron [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Updating instance_info_cache with network_info: [{"id": "ffee6438-b805-4a00-94a9-83fa6e39a230", "address": "fa:16:3e:ee:2b:32", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffee6438-b8", "ovs_interfaceid": "ffee6438-b805-4a00-94a9-83fa6e39a230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.668992] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159990, 'name': PowerOffVM_Task, 'duration_secs': 0.156646} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.669185] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 925.669445] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 925.670295] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed5acf5-2a8f-46fc-851b-10377081b07b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.689873] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 925.690527] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159991, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.690759] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e7e4ad0-84eb-44f4-bd06-a1d690ef4d56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.707878] env[61986]: DEBUG nova.objects.instance [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lazy-loading 'flavor' on Instance uuid a0fa2cbe-d8c3-462a-9744-aaa36a8d314e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.720943] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159992, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068296} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.721422] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.722327] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62dcea3e-7c2b-4298-8528-db0513aa3a4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.726900] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 925.727126] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 925.727305] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Deleting the datastore file [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.728249] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef7c6576-10d1-4f65-ae13-51869ad75595 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.752055] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/93c4f104-1812-4bb7-bfa7-cbf70a19ff51.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.753998] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2eaec488-6d57-4aa1-b4b9-80978490a58b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.768736] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 925.768736] env[61986]: value = "task-1159994" [ 925.768736] env[61986]: _type = "Task" [ 925.768736] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.777899] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 925.777899] env[61986]: value = "task-1159995" [ 925.777899] env[61986]: _type = "Task" [ 925.777899] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.781540] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159994, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.790359] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159995, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.870062] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1159989, 'name': CreateVM_Task, 'duration_secs': 0.512857} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.870062] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 925.870422] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.870508] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.870843] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.871144] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73a2188b-57d7-4148-99f7-172098d15272 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.876796] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 925.876796] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b1005b-ad04-c26f-66f0-9b3237a86e0d" [ 925.876796] env[61986]: _type = "Task" [ 925.876796] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.885971] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b1005b-ad04-c26f-66f0-9b3237a86e0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.947839] env[61986]: DEBUG nova.network.neutron [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Successfully updated port: 8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.081331] env[61986]: DEBUG nova.compute.utils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.082759] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.082939] env[61986]: DEBUG nova.network.neutron [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 926.105799] env[61986]: DEBUG oslo_concurrency.lockutils [req-ca455482-a501-4e4a-a079-54f75595e65b req-123fb022-d095-4eee-9941-9404c75fa8cb service nova] Releasing lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.127994] env[61986]: DEBUG nova.policy [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8562656b1e46628059ea24f9e16b26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0db18c866dc4ffb8bcb050f8ec6021d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 926.186564] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159991, 'name': CreateSnapshot_Task, 'duration_secs': 0.830886} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.190725] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 926.191798] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193ea784-7d99-47bc-8798-7cd5a98ff1fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.220634] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e2f9a981-50f4-4c0e-b6ae-ef7f9411e824 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.394s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.284274] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1159994, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099733} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.289732] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.289996] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 926.290280] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 926.299868] env[61986]: DEBUG nova.compute.manager [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.300280] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159995, 'name': ReconfigVM_Task, 'duration_secs': 0.489567} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.301102] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b78eab-8d27-4cf4-8c61-e3848050e672 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.305845] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/93c4f104-1812-4bb7-bfa7-cbf70a19ff51.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.311581] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c7a6019-41d9-4c8d-8f50-74d789eda9e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.327071] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 926.327071] env[61986]: value = "task-1159996" [ 926.327071] env[61986]: _type = "Task" [ 926.327071] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.069106] env[61986]: DEBUG nova.network.neutron [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Successfully created port: a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.069474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquiring lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.069474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquired lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.069549] env[61986]: DEBUG nova.network.neutron [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.070740] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.073382] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.073588] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.082967] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 927.087268] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.087741] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.087974] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.088116] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.088301] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.088442] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.088585] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.088781] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.088967] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.089168] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.089337] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.089520] env[61986]: DEBUG nova.virt.hardware [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.090309] env[61986]: INFO nova.compute.manager [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] instance snapshotting [ 927.093929] env[61986]: DEBUG nova.compute.manager [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Received event network-changed-8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.094130] env[61986]: DEBUG nova.compute.manager [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Refreshing instance network info cache due to event network-changed-8157a838-3c36-4936-a187-460e14d7f129. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 927.094324] env[61986]: DEBUG oslo_concurrency.lockutils [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] Acquiring lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.095477] env[61986]: INFO nova.compute.manager [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Detaching volume 753f1b3b-c85e-4cf3-83c8-a10d93fd03fd [ 927.103130] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-79e51de5-2361-459c-b317-f44755b89451 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.106540] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c827466e-443c-4d78-bd4b-d11ddb416199 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.109997] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01c3e98-da94-4f68-8b16-12bf805781f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.120705] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159996, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.120705] env[61986]: WARNING oslo_vmware.common.loopingcall [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] task run outlasted interval by 0.291872 sec [ 927.139799] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b1005b-ad04-c26f-66f0-9b3237a86e0d, 'name': SearchDatastore_Task, 'duration_secs': 0.010437} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.145814] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dc96da-f45d-4569-b495-3bf7cb8e8344 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.147680] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.147944] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.148203] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.148352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.148531] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.149083] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 927.149083] env[61986]: value = "task-1159997" [ 927.149083] env[61986]: _type = "Task" [ 927.149083] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.166673] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c047c47b-c6c4-4a53-8447-b115b1117d50 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.169702] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159996, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.176749] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ce07a4-3a7a-475b-8692-244b84d0a7ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.188150] env[61986]: INFO nova.virt.block_device [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Attempting to driver detach volume 753f1b3b-c85e-4cf3-83c8-a10d93fd03fd from mountpoint /dev/sdb [ 927.188521] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 927.188575] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252383', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'name': 'volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'serial': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 927.190819] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc38bc07-5242-4445-a401-16d1f3adb108 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.194401] env[61986]: DEBUG nova.network.neutron [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.199280] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.203027] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 927.207732] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44cf4e6c-b18a-4e8b-be19-a4737733351d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.214743] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159997, 'name': CloneVM_Task} progress is 21%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.239324] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efd1512-e508-4e58-93df-0fa2e4a9d7aa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.244402] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 927.244402] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528fbaac-fc69-0989-b7ee-cf421832065a" [ 927.244402] env[61986]: _type = "Task" [ 927.244402] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.253077] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51647230-f277-446c-a788-49e7b21aded2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.259669] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528fbaac-fc69-0989-b7ee-cf421832065a, 'name': SearchDatastore_Task, 'duration_secs': 0.011182} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.262387] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb1f9ad2-9062-4516-96a2-caccc203c84e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.289128] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5dbde0-5cdf-4cfa-85fc-a75dd9d1b937 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.295299] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 927.295299] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525db712-49b3-a901-0052-10feaefa6ebd" [ 927.295299] env[61986]: _type = "Task" [ 927.295299] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.312318] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] The volume has not been displaced from its original location: [datastore2] volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd/volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 927.319842] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 927.322384] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d5d84ad-0610-4807-b36b-a3243a5ce76e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.341589] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525db712-49b3-a901-0052-10feaefa6ebd, 'name': SearchDatastore_Task, 'duration_secs': 0.013684} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.345351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.345491] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3d5f4513-bbc4-404e-9d3e-340bd369fc3c/3d5f4513-bbc4-404e-9d3e-340bd369fc3c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 927.345830] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88578487-aa03-4f2e-912f-c197b2e1831c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.357348] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 927.357348] env[61986]: value = "task-1159998" [ 927.357348] env[61986]: _type = "Task" [ 927.357348] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.357348] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 927.357348] env[61986]: value = "task-1159999" [ 927.357348] env[61986]: _type = "Task" [ 927.357348] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.375959] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159998, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.376379] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.441018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaca8f69-ea82-4f37-8fc2-90cbaf546e53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.451376] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666ab698-4c98-4405-8dd5-2d808b688ff2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.484203] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801711c6-f848-4378-bba9-59dc1eaa5ff2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.493405] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e164f81b-65a2-423f-8270-b64d5e58a472 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.508784] env[61986]: DEBUG nova.compute.provider_tree [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.524226] env[61986]: DEBUG nova.network.neutron [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updating instance_info_cache with network_info: [{"id": "8157a838-3c36-4936-a187-460e14d7f129", "address": "fa:16:3e:c6:10:13", "network": {"id": "33c1c5c8-89b8-4be8-a7e6-6411f9c40a9c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1837194892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fe9c887c0d3485f8e6f01785b2a4743", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8157a838-3c", "ovs_interfaceid": "8157a838-3c36-4936-a187-460e14d7f129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.608066] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.643348] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159996, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.693077] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159997, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.731688] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.731971] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.732162] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.732357] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.732506] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.732660] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.732876] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.733156] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.733244] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.733412] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.733708] env[61986]: DEBUG nova.virt.hardware [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.734677] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 927.735543] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-143a97f3-4df3-4c69-b750-ee78e3dbc3e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.738322] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f924355d-cc79-44f8-bf47-6f6665b50ab3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.750207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ec04f7-932d-48c5-badb-1728165761a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.755144] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 927.755144] env[61986]: value = "task-1160000" [ 927.755144] env[61986]: _type = "Task" [ 927.755144] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.770506] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.776643] env[61986]: DEBUG oslo.service.loopingcall [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.777445] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 927.777755] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11518451-19b9-42a9-9532-8cd07f596279 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.795393] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160000, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.804643] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.804643] env[61986]: value = "task-1160001" [ 927.804643] env[61986]: _type = "Task" [ 927.804643] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.817430] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160001, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.873873] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159999, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.877519] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159998, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.013093] env[61986]: DEBUG nova.scheduler.client.report [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.029301] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Releasing lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.029617] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Instance network_info: |[{"id": "8157a838-3c36-4936-a187-460e14d7f129", "address": "fa:16:3e:c6:10:13", "network": {"id": "33c1c5c8-89b8-4be8-a7e6-6411f9c40a9c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1837194892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fe9c887c0d3485f8e6f01785b2a4743", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8157a838-3c", "ovs_interfaceid": "8157a838-3c36-4936-a187-460e14d7f129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.030932] env[61986]: DEBUG oslo_concurrency.lockutils [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] Acquired lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.031276] env[61986]: DEBUG nova.network.neutron [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Refreshing network info cache for port 8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 928.033037] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:10:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b89fd3b-0470-40c9-bb5b-d52c76c030e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8157a838-3c36-4936-a187-460e14d7f129', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.040271] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Creating folder: Project (8fe9c887c0d3485f8e6f01785b2a4743). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 928.041438] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af51d982-7b5e-4ced-869a-68570176bd1a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.059300] env[61986]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 928.059518] env[61986]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61986) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 928.060154] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Folder already exists: Project (8fe9c887c0d3485f8e6f01785b2a4743). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 928.060373] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Creating folder: Instances. Parent ref: group-v252357. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 928.060624] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9abaee27-1c8f-41d8-85ce-e6c1c1a2c0ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.073769] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Created folder: Instances in parent group-v252357. [ 928.074208] env[61986]: DEBUG oslo.service.loopingcall [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.074273] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 928.074453] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1ae6bb9-bd5d-4aab-932b-cb0e13f41ad3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.095320] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.095320] env[61986]: value = "task-1160004" [ 928.095320] env[61986]: _type = "Task" [ 928.095320] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.104140] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160004, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.110684] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.135176] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.135540] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.135717] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.135910] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.136099] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.136317] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.136668] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.136847] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.137090] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.137330] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.137652] env[61986]: DEBUG nova.virt.hardware [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.138769] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2742a47-173e-44e8-8471-0c4fc0751a85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.146300] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1159996, 'name': Rename_Task, 'duration_secs': 1.377799} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.147244] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.147552] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db255e88-e106-4c76-9b63-b2a553c49956 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.153687] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31311590-c3b5-4d7c-a337-e8a6e846d092 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.160594] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 928.160594] env[61986]: value = "task-1160005" [ 928.160594] env[61986]: _type = "Task" [ 928.160594] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.183024] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160005, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.190119] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159997, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.265799] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160000, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.315739] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160001, 'name': CreateVM_Task, 'duration_secs': 0.31631} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.315928] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 928.316722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.316883] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.317176] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.317464] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03e931db-0cbf-4b2d-a7c6-058bacbd5aee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.326217] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 928.326217] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a995a1-1889-6bd8-f5ed-64b1503559a9" [ 928.326217] env[61986]: _type = "Task" [ 928.326217] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.334811] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a995a1-1889-6bd8-f5ed-64b1503559a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.371169] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1159998, 'name': ReconfigVM_Task, 'duration_secs': 0.568915} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.371684] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.377069] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2251d0be-e8a0-47b2-a920-911ac031f097 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.392114] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1159999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524708} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.392899] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3d5f4513-bbc4-404e-9d3e-340bd369fc3c/3d5f4513-bbc4-404e-9d3e-340bd369fc3c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.393177] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.393869] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32c3bc07-b964-4b37-8dde-54c90007852f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.398943] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525a6f18-5ebc-39f7-a5e1-b0b5cc9d89e3/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 928.399960] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ae59a5-cba8-46e0-8c3c-6d2f567e9e48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.404695] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 928.404695] env[61986]: value = "task-1160006" [ 928.404695] env[61986]: _type = "Task" [ 928.404695] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.404882] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 928.404882] env[61986]: value = "task-1160007" [ 928.404882] env[61986]: _type = "Task" [ 928.404882] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.413191] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525a6f18-5ebc-39f7-a5e1-b0b5cc9d89e3/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 928.413375] env[61986]: ERROR oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525a6f18-5ebc-39f7-a5e1-b0b5cc9d89e3/disk-0.vmdk due to incomplete transfer. [ 928.414514] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4cb97c33-7787-4793-8134-ca130fbf373f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.419819] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160006, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.423152] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160007, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.424453] env[61986]: DEBUG oslo_vmware.rw_handles [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525a6f18-5ebc-39f7-a5e1-b0b5cc9d89e3/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 928.424648] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Uploaded image b36a707b-06ac-40f7-8b1a-e866a1c488a2 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 928.427289] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 928.427640] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f9ec1842-190e-4f11-b7cc-0b035b80d3f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.436869] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 928.436869] env[61986]: value = "task-1160008" [ 928.436869] env[61986]: _type = "Task" [ 928.436869] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.447566] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160008, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.517858] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.940s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.520385] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.634s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.522085] env[61986]: INFO nova.compute.claims [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.555302] env[61986]: INFO nova.scheduler.client.report [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Deleted allocations for instance a101dc3c-ca6b-4a72-a9b4-051b077a10fd [ 928.607475] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160004, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.677027] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160005, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.689343] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159997, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.766916] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160000, 'name': CreateSnapshot_Task, 'duration_secs': 0.948187} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.767572] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 928.768328] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebad381-fb68-4e03-a7d5-49e4e23332cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.838916] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a995a1-1889-6bd8-f5ed-64b1503559a9, 'name': SearchDatastore_Task, 'duration_secs': 0.022484} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.841766] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.841930] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.842236] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.843229] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.843229] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.843229] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a640e84-443d-4294-b6bc-fa8d109c5e50 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.853594] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.853594] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 928.854663] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76b72e26-ad3e-4991-9007-49074b7a0a7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.862364] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 928.862364] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527a564f-e434-8abb-2501-2f7958ff854f" [ 928.862364] env[61986]: _type = "Task" [ 928.862364] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.876828] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527a564f-e434-8abb-2501-2f7958ff854f, 'name': SearchDatastore_Task, 'duration_secs': 0.010406} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.880820] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a43f536-c0b0-45ed-861b-25f51351313a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.889369] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 928.889369] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c29ae4-fae4-c1ee-50f5-de57e6857969" [ 928.889369] env[61986]: _type = "Task" [ 928.889369] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.904481] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c29ae4-fae4-c1ee-50f5-de57e6857969, 'name': SearchDatastore_Task} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.905029] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.905491] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 928.906095] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b60918b-4bbb-4d26-b48b-4e93f8b63e58 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.919337] env[61986]: DEBUG nova.network.neutron [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updated VIF entry in instance network info cache for port 8157a838-3c36-4936-a187-460e14d7f129. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 928.919602] env[61986]: DEBUG nova.network.neutron [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updating instance_info_cache with network_info: [{"id": "8157a838-3c36-4936-a187-460e14d7f129", "address": "fa:16:3e:c6:10:13", "network": {"id": "33c1c5c8-89b8-4be8-a7e6-6411f9c40a9c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1837194892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fe9c887c0d3485f8e6f01785b2a4743", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8157a838-3c", "ovs_interfaceid": "8157a838-3c36-4936-a187-460e14d7f129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.923801] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 928.923801] env[61986]: value = "task-1160009" [ 928.923801] env[61986]: _type = "Task" [ 928.923801] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.931293] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160007, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076537} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.931293] env[61986]: DEBUG oslo_vmware.api [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160006, 'name': ReconfigVM_Task, 'duration_secs': 0.303948} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.934237] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.936671] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252383', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'name': 'volume-753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd', 'serial': '753f1b3b-c85e-4cf3-83c8-a10d93fd03fd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 928.938363] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53537ba-568d-40a4-addf-e5cb6a623c24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.948710] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.968875] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 3d5f4513-bbc4-404e-9d3e-340bd369fc3c/3d5f4513-bbc4-404e-9d3e-340bd369fc3c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.972531] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccc503f0-ed4d-48d4-bca1-4ce4f9665040 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.987163] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160008, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.995042] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 928.995042] env[61986]: value = "task-1160010" [ 928.995042] env[61986]: _type = "Task" [ 928.995042] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.005044] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160010, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.036023] env[61986]: DEBUG nova.network.neutron [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Successfully updated port: a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.066149] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2116bb0a-6a5b-4bb1-9d8c-308f438e135f tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.334s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.067166] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 18.815s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.067392] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.067591] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.067754] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.069910] env[61986]: INFO nova.compute.manager [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Terminating instance [ 929.071583] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.071752] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquired lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.071907] env[61986]: DEBUG nova.network.neutron [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.107238] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160004, 'name': CreateVM_Task, 'duration_secs': 0.584246} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.107652] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 929.108187] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252362', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'name': 'volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5170aa51-3307-42b1-b0dd-645dd4036e5b', 'attached_at': '', 'detached_at': '', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'serial': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5'}, 'guest_format': None, 'attachment_id': '14b65da0-560d-4c32-8e2d-b0fd271f1ab4', 'boot_index': 0, 'disk_bus': None, 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61986) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 929.108405] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Root volume attach. Driver type: vmdk {{(pid=61986) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 929.109336] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c08ce5-2b95-4a03-a554-71d37b4d197c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.118576] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1477cf9-f176-4e1e-85aa-a5fcf3683592 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.128276] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c201e3-74c6-4ed0-a373-9056b70b15d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.140022] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-20c00c7d-1209-4051-97d3-c2f7e2addb41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.147190] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 929.147190] env[61986]: value = "task-1160011" [ 929.147190] env[61986]: _type = "Task" [ 929.147190] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.158047] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.160642] env[61986]: DEBUG nova.compute.manager [req-547e3d62-bf16-4bd4-81fa-40089c162bc5 req-4241950f-c04c-4f7f-9163-361a5191b33c service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Received event network-vif-plugged-a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.161551] env[61986]: DEBUG oslo_concurrency.lockutils [req-547e3d62-bf16-4bd4-81fa-40089c162bc5 req-4241950f-c04c-4f7f-9163-361a5191b33c service nova] Acquiring lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.161551] env[61986]: DEBUG oslo_concurrency.lockutils [req-547e3d62-bf16-4bd4-81fa-40089c162bc5 req-4241950f-c04c-4f7f-9163-361a5191b33c service nova] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.161551] env[61986]: DEBUG oslo_concurrency.lockutils [req-547e3d62-bf16-4bd4-81fa-40089c162bc5 req-4241950f-c04c-4f7f-9163-361a5191b33c service nova] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.161551] env[61986]: DEBUG nova.compute.manager [req-547e3d62-bf16-4bd4-81fa-40089c162bc5 req-4241950f-c04c-4f7f-9163-361a5191b33c service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] No waiting events found dispatching network-vif-plugged-a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.161551] env[61986]: WARNING nova.compute.manager [req-547e3d62-bf16-4bd4-81fa-40089c162bc5 req-4241950f-c04c-4f7f-9163-361a5191b33c service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Received unexpected event network-vif-plugged-a9804290-f49e-46f6-90c6-3ce3e529d094 for instance with vm_state building and task_state spawning. [ 929.173664] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160005, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.187361] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1159997, 'name': CloneVM_Task, 'duration_secs': 1.873473} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.187706] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Created linked-clone VM from snapshot [ 929.188592] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8dc1ea-0a04-4fcb-b30f-ea388f1dba7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.198068] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Uploading image 95fbc9d9-8c18-4590-bfb5-f7dc569e3007 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 929.210637] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 929.210948] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b688e211-6e9a-4431-b078-263e7999824a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.219375] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 929.219375] env[61986]: value = "task-1160012" [ 929.219375] env[61986]: _type = "Task" [ 929.219375] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.229463] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160012, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.244880] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "98337bb6-9502-4d4c-af00-028659b246bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.245164] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "98337bb6-9502-4d4c-af00-028659b246bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.245409] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "98337bb6-9502-4d4c-af00-028659b246bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.245609] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "98337bb6-9502-4d4c-af00-028659b246bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.245777] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "98337bb6-9502-4d4c-af00-028659b246bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.248532] env[61986]: INFO nova.compute.manager [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Terminating instance [ 929.252080] env[61986]: DEBUG nova.compute.manager [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.253027] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d31e3bae-6e33-4c7c-9030-0e1df9ab4185 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.263920] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5a3a95-7b7e-4231-ad22-e23c15b8724d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.292167] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 929.306350] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-201a2d5d-e244-49a1-b666-be71cbfad13b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.310613] env[61986]: WARNING nova.virt.vmwareapi.driver [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 98337bb6-9502-4d4c-af00-028659b246bf could not be found. [ 929.310845] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 929.311397] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a29f7d5-00a5-47d5-8df7-0075157a5e44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.324341] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0ce60b-d516-4ac5-abbe-a724e26c14cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.335648] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 929.335648] env[61986]: value = "task-1160013" [ 929.335648] env[61986]: _type = "Task" [ 929.335648] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.349598] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.377953] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 98337bb6-9502-4d4c-af00-028659b246bf could not be found. [ 929.378361] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 929.379015] env[61986]: INFO nova.compute.manager [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Took 0.13 seconds to destroy the instance on the hypervisor. [ 929.379424] env[61986]: DEBUG oslo.service.loopingcall [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.380249] env[61986]: DEBUG nova.compute.manager [-] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.380439] env[61986]: DEBUG nova.network.neutron [-] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 929.423966] env[61986]: DEBUG oslo_concurrency.lockutils [req-8ae96b99-eb52-4ce5-a854-b6c2e6c08406 req-4c200ea4-767f-47cb-9bf5-66f66f46d426 service nova] Releasing lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.444584] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160009, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.458291] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160008, 'name': Destroy_Task, 'duration_secs': 0.745638} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.458788] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroyed the VM [ 929.459315] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 929.459770] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d038f8d8-4bad-4acc-9343-9aa8b41c6812 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.469189] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 929.469189] env[61986]: value = "task-1160014" [ 929.469189] env[61986]: _type = "Task" [ 929.469189] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.482986] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.495475] env[61986]: DEBUG nova.objects.instance [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lazy-loading 'flavor' on Instance uuid a0fa2cbe-d8c3-462a-9744-aaa36a8d314e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.509045] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160010, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.542463] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.542711] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.542967] env[61986]: DEBUG nova.network.neutron [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.575717] env[61986]: DEBUG nova.compute.utils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Can not refresh info_cache because instance was not found {{(pid=61986) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 929.606152] env[61986]: DEBUG nova.network.neutron [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.657381] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 38%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.676749] env[61986]: DEBUG oslo_vmware.api [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160005, 'name': PowerOnVM_Task, 'duration_secs': 1.266289} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.676749] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.680069] env[61986]: INFO nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Took 10.13 seconds to spawn the instance on the hypervisor. [ 929.680069] env[61986]: DEBUG nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.682206] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2451252b-807b-4bcb-8590-71fe78c1f6b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.733786] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160012, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.828288] env[61986]: DEBUG nova.network.neutron [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.851213] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 93%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.943307] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557547} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.946659] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 929.947194] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.947804] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98ec9244-58d3-4066-954e-15cad52e3a26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.955197] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 929.955197] env[61986]: value = "task-1160015" [ 929.955197] env[61986]: _type = "Task" [ 929.955197] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.964820] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.978671] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.006599] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160010, 'name': ReconfigVM_Task, 'duration_secs': 0.670585} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.008325] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 3d5f4513-bbc4-404e-9d3e-340bd369fc3c/3d5f4513-bbc4-404e-9d3e-340bd369fc3c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.011694] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56e3cf93-a156-4211-922f-c75c3ee2402c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.019276] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 930.019276] env[61986]: value = "task-1160016" [ 930.019276] env[61986]: _type = "Task" [ 930.019276] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.028664] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160016, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.081690] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73f7153-2cd8-4bd8-964e-12b795c08cd0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.090026] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8720ec81-4bf4-43b5-9747-1157bfb89c89 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.123324] env[61986]: DEBUG nova.network.neutron [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 930.126162] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae56c59-b766-41aa-9202-eed956ec8162 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.136412] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c00de56-09c7-4573-8c6c-ed6b06768895 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.150867] env[61986]: DEBUG nova.compute.provider_tree [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.164307] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 53%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.206798] env[61986]: INFO nova.compute.manager [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Took 36.08 seconds to build instance. [ 930.235168] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160012, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.312819] env[61986]: DEBUG nova.network.neutron [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Updating instance_info_cache with network_info: [{"id": "a9804290-f49e-46f6-90c6-3ce3e529d094", "address": "fa:16:3e:e5:e6:a1", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9804290-f4", "ovs_interfaceid": "a9804290-f49e-46f6-90c6-3ce3e529d094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.334185] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Releasing lock "refresh_cache-a101dc3c-ca6b-4a72-a9b4-051b077a10fd" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.334185] env[61986]: DEBUG nova.compute.manager [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.334185] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.334185] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79e76070-218a-4daa-a794-b7f474807bba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.346149] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c8b3bd-0a0a-45bd-9018-85259a6b698a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.363364] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.391500] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a101dc3c-ca6b-4a72-a9b4-051b077a10fd could not be found. [ 930.391720] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 930.391911] env[61986]: INFO nova.compute.manager [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Took 0.06 seconds to destroy the instance on the hypervisor. [ 930.392188] env[61986]: DEBUG oslo.service.loopingcall [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.392496] env[61986]: DEBUG nova.compute.manager [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 930.392595] env[61986]: DEBUG nova.network.neutron [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 930.468164] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09142} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.468164] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.468837] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98719d37-c707-4139-931a-883598a66a2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.485214] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.507326] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.507326] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b1a9df5-5df1-4d1b-8e9a-5251846c268c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.520312] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fa1afa6e-01df-419e-b8f2-cd5d6afbe0c6 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.446s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.531426] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160016, 'name': Rename_Task, 'duration_secs': 0.16464} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.531719] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.531978] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fdeef603-40c9-4982-845d-cdf9b7f63391 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.536096] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 930.536096] env[61986]: value = "task-1160017" [ 930.536096] env[61986]: _type = "Task" [ 930.536096] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.541503] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 930.541503] env[61986]: value = "task-1160018" [ 930.541503] env[61986]: _type = "Task" [ 930.541503] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.549339] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.555656] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160018, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.638714] env[61986]: DEBUG nova.network.neutron [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 930.641749] env[61986]: DEBUG nova.network.neutron [-] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.659244] env[61986]: DEBUG nova.scheduler.client.report [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.669083] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 65%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.711119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f54eb1d4-4d3f-406c-b4e5-696bfbe2faa1 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.205s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.739792] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160012, 'name': Destroy_Task, 'duration_secs': 1.129462} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.740124] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Destroyed the VM [ 930.740443] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 930.740804] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a42428c9-1552-4369-b285-7150dc2ed7af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.748981] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 930.748981] env[61986]: value = "task-1160019" [ 930.748981] env[61986]: _type = "Task" [ 930.748981] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.766476] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160019, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.768195] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "86c9e45c-b3f1-4004-9acc-190b11a4a926" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.768534] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.768821] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "86c9e45c-b3f1-4004-9acc-190b11a4a926-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.769112] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.769352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.772360] env[61986]: INFO nova.compute.manager [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Terminating instance [ 930.775104] env[61986]: DEBUG nova.compute.manager [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.775390] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.776749] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f03770-4cbf-4db6-aa56-e178f5b42c39 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.788319] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 930.788792] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1540573d-17e1-43fa-aa2d-0d85e4fd0613 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.802855] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 930.802855] env[61986]: value = "task-1160020" [ 930.802855] env[61986]: _type = "Task" [ 930.802855] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.813839] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1160020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.816566] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.817059] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Instance network_info: |[{"id": "a9804290-f49e-46f6-90c6-3ce3e529d094", "address": "fa:16:3e:e5:e6:a1", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9804290-f4", "ovs_interfaceid": "a9804290-f49e-46f6-90c6-3ce3e529d094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.817736] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:e6:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9804290-f49e-46f6-90c6-3ce3e529d094', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.831370] env[61986]: DEBUG oslo.service.loopingcall [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.833916] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.833916] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b95122f-7957-490e-8f37-f2fa4ec6b7ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.870057] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.872049] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.872049] env[61986]: value = "task-1160021" [ 930.872049] env[61986]: _type = "Task" [ 930.872049] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.883615] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160021, 'name': CreateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.938587] env[61986]: INFO nova.compute.manager [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Rescuing [ 930.939071] env[61986]: DEBUG oslo_concurrency.lockutils [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.939301] env[61986]: DEBUG oslo_concurrency.lockutils [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.939496] env[61986]: DEBUG nova.network.neutron [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 930.982696] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.047091] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.060494] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160018, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.146747] env[61986]: DEBUG nova.network.neutron [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.148582] env[61986]: INFO nova.compute.manager [-] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Took 1.77 seconds to deallocate network for instance. [ 931.169468] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 78%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.171259] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.651s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.173393] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.175245] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 19.101s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.263603] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160019, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.313783] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1160020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.372611] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.384199] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160021, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.449131] env[61986]: DEBUG nova.compute.manager [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Received event network-changed-a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 931.449436] env[61986]: DEBUG nova.compute.manager [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Refreshing instance network info cache due to event network-changed-a9804290-f49e-46f6-90c6-3ce3e529d094. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 931.449734] env[61986]: DEBUG oslo_concurrency.lockutils [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] Acquiring lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.450014] env[61986]: DEBUG oslo_concurrency.lockutils [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] Acquired lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.450357] env[61986]: DEBUG nova.network.neutron [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Refreshing network info cache for port a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 931.485714] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.547014] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160017, 'name': ReconfigVM_Task, 'duration_secs': 0.587089} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.550415] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Reconfigured VM instance instance-0000004b to attach disk [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452/dea128a9-aed0-40b2-ae17-c068ea8e3452.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.551145] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abde5f93-b109-4502-8190-52ed395c42ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.558303] env[61986]: DEBUG oslo_vmware.api [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160018, 'name': PowerOnVM_Task, 'duration_secs': 0.566949} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.560048] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.561057] env[61986]: INFO nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Took 9.29 seconds to spawn the instance on the hypervisor. [ 931.561057] env[61986]: DEBUG nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.561057] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 931.561057] env[61986]: value = "task-1160022" [ 931.561057] env[61986]: _type = "Task" [ 931.561057] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.564835] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12ca702-924c-4922-84b8-fcdd79a99231 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.576872] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160022, 'name': Rename_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.654143] env[61986]: INFO nova.compute.manager [-] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Took 1.26 seconds to deallocate network for instance. [ 931.672821] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 92%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.679712] env[61986]: DEBUG nova.compute.utils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.682293] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.682585] env[61986]: DEBUG nova.network.neutron [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.688055] env[61986]: INFO nova.compute.claims [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.725665] env[61986]: INFO nova.compute.manager [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Took 0.57 seconds to detach 1 volumes for instance. [ 931.726180] env[61986]: DEBUG nova.compute.manager [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Deleting volume: 6d6d6476-0f93-4404-b58d-3092cea80b1b {{(pid=61986) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 931.763252] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160019, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.811758] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1160020, 'name': PowerOffVM_Task, 'duration_secs': 0.63659} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.815023] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 931.815023] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 931.816568] env[61986]: DEBUG nova.policy [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e8ff762d1dc4901b72b9a25be4ef39f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cdbc87d45204921a43ac1555391b54a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 931.819420] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed2ea03b-bb22-41c8-bb6f-a5269a3cf540 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.867631] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.868117] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.868419] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.883207] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160021, 'name': CreateVM_Task, 'duration_secs': 0.534521} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.883347] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 931.884063] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.884254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.884608] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.884881] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a395c0b-fd0c-4132-9a15-a18578c35185 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.890523] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 931.890523] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b92230-afb7-6e1d-cb9e-24a78e856725" [ 931.890523] env[61986]: _type = "Task" [ 931.890523] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.899160] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b92230-afb7-6e1d-cb9e-24a78e856725, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.931525] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 931.931752] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 931.931938] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Deleting the datastore file [datastore2] 86c9e45c-b3f1-4004-9acc-190b11a4a926 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.932242] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-578d3122-6578-49f2-a85f-ef080ce51fd2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.939373] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for the task: (returnval){ [ 931.939373] env[61986]: value = "task-1160025" [ 931.939373] env[61986]: _type = "Task" [ 931.939373] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.948225] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1160025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.956483] env[61986]: DEBUG nova.network.neutron [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updating instance_info_cache with network_info: [{"id": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "address": "fa:16:3e:41:aa:c3", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4411312-86", "ovs_interfaceid": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.983799] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.079270] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160022, 'name': Rename_Task, 'duration_secs': 0.24614} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.079635] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 932.080028] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a12cc15e-c5ec-42aa-ba6a-23bc5595dadf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.092236] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 932.092236] env[61986]: value = "task-1160026" [ 932.092236] env[61986]: _type = "Task" [ 932.092236] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.099024] env[61986]: INFO nova.compute.manager [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Took 33.01 seconds to build instance. [ 932.110298] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160026, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.129397] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] Acquiring lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.129578] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] Acquired lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.129755] env[61986]: DEBUG nova.network.neutron [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 932.161318] env[61986]: INFO nova.compute.manager [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance disappeared during terminate [ 932.161632] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dea8b993-ab7b-4e6a-b6cc-4094d9753d48 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "a101dc3c-ca6b-4a72-a9b4-051b077a10fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.094s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.170481] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 97%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.182848] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.197401] env[61986]: INFO nova.compute.resource_tracker [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating resource usage from migration b772f020-d14b-4ce6-a2c7-eb4d645079b2 [ 932.273581] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160019, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.280797] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.356636] env[61986]: DEBUG nova.network.neutron [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Successfully created port: 514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.360798] env[61986]: DEBUG nova.network.neutron [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Updated VIF entry in instance network info cache for port a9804290-f49e-46f6-90c6-3ce3e529d094. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 932.361554] env[61986]: DEBUG nova.network.neutron [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Updating instance_info_cache with network_info: [{"id": "a9804290-f49e-46f6-90c6-3ce3e529d094", "address": "fa:16:3e:e5:e6:a1", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9804290-f4", "ovs_interfaceid": "a9804290-f49e-46f6-90c6-3ce3e529d094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.373217] env[61986]: INFO nova.compute.manager [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Detaching volume 47fff893-4caa-4a2a-ba29-4ee2766f6406 [ 932.380859] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.408021] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b92230-afb7-6e1d-cb9e-24a78e856725, 'name': SearchDatastore_Task, 'duration_secs': 0.01474} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.408021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.408021] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.408021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.408021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.408021] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.408922] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afc76ec0-e8b9-4bbd-a732-767fe8969c41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.424026] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.424026] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 932.424026] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67caf444-2ac2-4df7-87b3-1893cce93e02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.431617] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 932.431617] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a7bbc9-c52e-de92-a626-610f47ee88f2" [ 932.431617] env[61986]: _type = "Task" [ 932.431617] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.438903] env[61986]: INFO nova.virt.block_device [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Attempting to driver detach volume 47fff893-4caa-4a2a-ba29-4ee2766f6406 from mountpoint /dev/sdc [ 932.439344] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 932.439830] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252394', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'name': 'volume-47fff893-4caa-4a2a-ba29-4ee2766f6406', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'serial': '47fff893-4caa-4a2a-ba29-4ee2766f6406'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 932.441642] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c80d97e-a8c0-4ac3-a0c2-b1da1acf877b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.451706] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a7bbc9-c52e-de92-a626-610f47ee88f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.484673] env[61986]: DEBUG oslo_concurrency.lockutils [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.487053] env[61986]: DEBUG oslo_vmware.api [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Task: {'id': task-1160025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.39151} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.495149] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77be5640-a43b-4372-bf64-7b5f401aa0d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.496769] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.497030] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 932.497189] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 932.497377] env[61986]: INFO nova.compute.manager [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Took 1.72 seconds to destroy the instance on the hypervisor. [ 932.497715] env[61986]: DEBUG oslo.service.loopingcall [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.498680] env[61986]: DEBUG nova.compute.manager [-] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 932.498801] env[61986]: DEBUG nova.network.neutron [-] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 932.506285] env[61986]: DEBUG oslo_vmware.api [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160014, 'name': RemoveSnapshot_Task, 'duration_secs': 2.995218} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.508963] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 932.508963] env[61986]: INFO nova.compute.manager [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 17.26 seconds to snapshot the instance on the hypervisor. [ 932.511734] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95de9761-c87d-4973-bd83-02bc5fa04fe8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.545147] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ae0217-a7f7-4708-9d30-cec40e86971b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.562840] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] The volume has not been displaced from its original location: [datastore1] volume-47fff893-4caa-4a2a-ba29-4ee2766f6406/volume-47fff893-4caa-4a2a-ba29-4ee2766f6406.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 932.568479] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfiguring VM instance instance-0000003a to detach disk 2002 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 932.571660] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ebaff65-1c33-4a6e-adc4-1eab9053a98f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.590466] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 932.590466] env[61986]: value = "task-1160027" [ 932.590466] env[61986]: _type = "Task" [ 932.590466] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.600748] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160027, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.605997] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb17c3b-592c-41d8-a14a-184f7be7d9f0 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.845s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.611385] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160026, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.669504] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task} progress is 98%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.767484] env[61986]: DEBUG oslo_vmware.api [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160019, 'name': RemoveSnapshot_Task, 'duration_secs': 1.811679} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.767781] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 932.806509] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81f0260-8849-485b-8451-266e95550a33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.815966] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbab6139-51d0-4bc6-9453-b3ec409351d2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.854622] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285afa61-b676-4e2f-9171-2d98abe5744f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.865879] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04063e2f-c043-4416-b3dc-1ca71ba22793 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.870087] env[61986]: DEBUG oslo_concurrency.lockutils [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] Releasing lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.870350] env[61986]: DEBUG nova.compute.manager [req-b8a7e931-0118-4250-888e-5dc05c9228a6 req-16191eeb-50de-4ac5-afe9-a812b92b15bb service nova] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Received event network-vif-deleted-494be155-c4b7-446d-b2ca-d49f22c7cc85 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.875589] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160013, 'name': CloneVM_Task, 'duration_secs': 3.473675} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.884059] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Created linked-clone VM from snapshot [ 932.884778] env[61986]: DEBUG nova.compute.provider_tree [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.890121] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0481a7c-eba2-44a8-a39c-70f3259785e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.898865] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Uploading image 600544b5-22c3-4141-8687-1b96be17aca7 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 932.926704] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 932.926704] env[61986]: value = "vm-252403" [ 932.926704] env[61986]: _type = "VirtualMachine" [ 932.926704] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 932.927032] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-daf512ac-3efc-4c58-b64d-72723d24ad71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.934926] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lease: (returnval){ [ 932.934926] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709bd6-98d4-f4de-58fd-b21440fc6a9a" [ 932.934926] env[61986]: _type = "HttpNfcLease" [ 932.934926] env[61986]: } obtained for exporting VM: (result){ [ 932.934926] env[61986]: value = "vm-252403" [ 932.934926] env[61986]: _type = "VirtualMachine" [ 932.934926] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 932.935385] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the lease: (returnval){ [ 932.935385] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709bd6-98d4-f4de-58fd-b21440fc6a9a" [ 932.935385] env[61986]: _type = "HttpNfcLease" [ 932.935385] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 932.946185] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a7bbc9-c52e-de92-a626-610f47ee88f2, 'name': SearchDatastore_Task, 'duration_secs': 0.015624} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.946185] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d306362a-618d-4d1d-9e59-3726450fa180 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.950323] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.950323] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709bd6-98d4-f4de-58fd-b21440fc6a9a" [ 932.950323] env[61986]: _type = "HttpNfcLease" [ 932.950323] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 932.952476] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 932.952476] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529d5668-a49f-b2e2-25cb-ab9797076afb" [ 932.952476] env[61986]: _type = "Task" [ 932.952476] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.960392] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529d5668-a49f-b2e2-25cb-ab9797076afb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.020763] env[61986]: DEBUG nova.network.neutron [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Updating instance_info_cache with network_info: [{"id": "ffee6438-b805-4a00-94a9-83fa6e39a230", "address": "fa:16:3e:ee:2b:32", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffee6438-b8", "ovs_interfaceid": "ffee6438-b805-4a00-94a9-83fa6e39a230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.051631] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.051899] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fded75b-1f56-4995-a1e2-0871ea6958d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.062867] env[61986]: DEBUG nova.compute.manager [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Found 3 images (rotation: 2) {{(pid=61986) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 933.063111] env[61986]: DEBUG nova.compute.manager [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Rotating out 1 backups {{(pid=61986) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 933.063269] env[61986]: DEBUG nova.compute.manager [None req-be86040d-1909-4f8f-80f6-778b01a5a229 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleting image 46c39e0f-d3b3-49a0-8ab7-9754db5273df {{(pid=61986) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 933.066126] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 933.066126] env[61986]: value = "task-1160029" [ 933.066126] env[61986]: _type = "Task" [ 933.066126] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.074802] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160029, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.099913] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160027, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.109351] env[61986]: DEBUG oslo_vmware.api [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160026, 'name': PowerOnVM_Task, 'duration_secs': 0.741658} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.109608] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 933.109828] env[61986]: DEBUG nova.compute.manager [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.111050] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a2e701-1739-4f3f-aa7c-22f59805781d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.170710] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160011, 'name': RelocateVM_Task, 'duration_secs': 3.662461} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.171093] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 933.171281] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252362', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'name': 'volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5170aa51-3307-42b1-b0dd-645dd4036e5b', 'attached_at': '', 'detached_at': '', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'serial': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 933.172191] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769d7d8b-47bc-4555-bec6-c1c917ebfe40 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.193393] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3981eef6-18e1-493c-997f-0dfb15066e98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.196935] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.223329] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5/volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.225821] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08eff18e-8167-43cc-8e5c-b2cdf3facd5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.249097] env[61986]: DEBUG nova.virt.hardware [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.249655] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45dcf6f-6fa9-4e36-afed-a464e3f89a27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.253124] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 933.253124] env[61986]: value = "task-1160030" [ 933.253124] env[61986]: _type = "Task" [ 933.253124] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.260041] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80f44af-afac-43c6-8363-d9ff6ea86e0f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.266804] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.276848] env[61986]: WARNING nova.compute.manager [None req-18073b38-7ec4-44a4-9ead-78d8e7183228 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Image not found during snapshot: nova.exception.ImageNotFound: Image 95fbc9d9-8c18-4590-bfb5-f7dc569e3007 could not be found. [ 933.394936] env[61986]: DEBUG nova.scheduler.client.report [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.445886] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 933.445886] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709bd6-98d4-f4de-58fd-b21440fc6a9a" [ 933.445886] env[61986]: _type = "HttpNfcLease" [ 933.445886] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 933.446966] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 933.446966] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709bd6-98d4-f4de-58fd-b21440fc6a9a" [ 933.446966] env[61986]: _type = "HttpNfcLease" [ 933.446966] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 933.446966] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cae63d-bd88-48e7-a0b1-5788aa424e4d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.454863] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52714534-8578-5599-14e2-ae546d642966/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 933.455066] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52714534-8578-5599-14e2-ae546d642966/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 933.532204] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] Releasing lock "refresh_cache-3d5f4513-bbc4-404e-9d3e-340bd369fc3c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.532406] env[61986]: DEBUG nova.compute.manager [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Inject network info {{(pid=61986) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 933.532743] env[61986]: DEBUG nova.compute.manager [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] network_info to inject: |[{"id": "ffee6438-b805-4a00-94a9-83fa6e39a230", "address": "fa:16:3e:ee:2b:32", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffee6438-b8", "ovs_interfaceid": "ffee6438-b805-4a00-94a9-83fa6e39a230", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 933.540113] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Reconfiguring VM instance to set the machine id {{(pid=61986) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 933.546087] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0822c3ff-a61b-4390-995d-430f2e6871a7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.555612] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529d5668-a49f-b2e2-25cb-ab9797076afb, 'name': SearchDatastore_Task, 'duration_secs': 0.032365} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.556224] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.556488] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] f15bf19d-b86a-4b0a-ac1a-9df8e77e9382/f15bf19d-b86a-4b0a-ac1a-9df8e77e9382.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 933.559159] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8dcbbfc-475b-4b2c-8f0d-f89f0035d683 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.562316] env[61986]: DEBUG oslo_vmware.api [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] Waiting for the task: (returnval){ [ 933.562316] env[61986]: value = "task-1160031" [ 933.562316] env[61986]: _type = "Task" [ 933.562316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.568017] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 933.568017] env[61986]: value = "task-1160032" [ 933.568017] env[61986]: _type = "Task" [ 933.568017] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.575585] env[61986]: DEBUG oslo_vmware.api [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] Task: {'id': task-1160031, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.579748] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160029, 'name': PowerOffVM_Task, 'duration_secs': 0.361561} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.586047] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 933.586047] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.586047] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a7d5af-e143-48e6-a6d0-874f6acaf4cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.609258] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3202b840-a344-464d-9ab1-e29096b97a18 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.613996] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fadeb374-5a54-4a61-9f6e-f3e5f450c504 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.619062] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160027, 'name': ReconfigVM_Task, 'duration_secs': 0.511452} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.622327] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Reconfigured VM instance instance-0000003a to detach disk 2002 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.635728] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1ef8f8b-f090-4648-b66a-a92018a09869 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.646809] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.651876] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 933.651876] env[61986]: value = "task-1160033" [ 933.651876] env[61986]: _type = "Task" [ 933.651876] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.659578] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160033, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.666123] env[61986]: DEBUG nova.network.neutron [-] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.669194] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.670774] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4c6ac28-fb6a-4fa9-9dbe-b96bd1879913 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.679033] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 933.679033] env[61986]: value = "task-1160034" [ 933.679033] env[61986]: _type = "Task" [ 933.679033] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.693092] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 933.693320] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.693563] env[61986]: DEBUG oslo_concurrency.lockutils [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.693710] env[61986]: DEBUG oslo_concurrency.lockutils [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.693984] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.694194] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1e40a9f-0b9d-4e20-ad76-448e1e57f865 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.703623] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.703856] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 933.704808] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ebf0358-7be4-443c-8569-93af2d53c5a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.715071] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 933.715071] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5282b037-d456-0296-755e-b89f8d454e34" [ 933.715071] env[61986]: _type = "Task" [ 933.715071] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.720131] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5282b037-d456-0296-755e-b89f8d454e34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.763699] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160030, 'name': ReconfigVM_Task, 'duration_secs': 0.462755} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.764265] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Reconfigured VM instance instance-0000004e to attach disk [datastore1] volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5/volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.770109] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e802ec1c-2ad0-4ef0-ac2d-0e1aef8edc60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.790523] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 933.790523] env[61986]: value = "task-1160035" [ 933.790523] env[61986]: _type = "Task" [ 933.790523] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.791613] env[61986]: DEBUG nova.compute.manager [req-e4622224-6f14-4d0a-9793-56b086aa6997 req-547b9a21-8935-40a1-9aa6-4545d1fa26e1 service nova] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Received event network-vif-deleted-812b5c34-903f-47de-b3f8-649b3a84eb27 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.805183] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160035, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.900676] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.725s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.902989] env[61986]: INFO nova.compute.manager [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Migrating [ 933.910252] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.245s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.910696] env[61986]: DEBUG nova.objects.instance [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lazy-loading 'resources' on Instance uuid 6f9349b7-886a-4077-8f6d-a9800ab353fe {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.078119] env[61986]: DEBUG oslo_vmware.api [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] Task: {'id': task-1160031, 'name': ReconfigVM_Task, 'duration_secs': 0.176482} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.079425] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5500b41f-7262-4efa-8d4c-d2b90cc60fcc tempest-ServersAdminTestJSON-580684259 tempest-ServersAdminTestJSON-580684259-project-admin] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Reconfigured VM instance to set the machine id {{(pid=61986) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 934.083848] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160032, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.162266] env[61986]: DEBUG oslo_vmware.api [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160033, 'name': ReconfigVM_Task, 'duration_secs': 0.160191} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.162724] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252394', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'name': 'volume-47fff893-4caa-4a2a-ba29-4ee2766f6406', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0fa2cbe-d8c3-462a-9744-aaa36a8d314e', 'attached_at': '', 'detached_at': '', 'volume_id': '47fff893-4caa-4a2a-ba29-4ee2766f6406', 'serial': '47fff893-4caa-4a2a-ba29-4ee2766f6406'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 934.174232] env[61986]: INFO nova.compute.manager [-] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Took 1.67 seconds to deallocate network for instance. [ 934.231543] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5282b037-d456-0296-755e-b89f8d454e34, 'name': SearchDatastore_Task, 'duration_secs': 0.011376} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.232564] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e128395-33d8-4a93-9949-b07e29c31aa3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.238755] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 934.238755] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52add07b-016f-f9c1-931f-c3a1ecb6fb5a" [ 934.238755] env[61986]: _type = "Task" [ 934.238755] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.248206] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52add07b-016f-f9c1-931f-c3a1ecb6fb5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.308086] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160035, 'name': ReconfigVM_Task, 'duration_secs': 0.182115} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.308814] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252362', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'name': 'volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5170aa51-3307-42b1-b0dd-645dd4036e5b', 'attached_at': '', 'detached_at': '', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'serial': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 934.311741] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9bc7807d-7875-42eb-a956-5439a0fa035b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.316784] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 934.316784] env[61986]: value = "task-1160036" [ 934.316784] env[61986]: _type = "Task" [ 934.316784] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.326173] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160036, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.352644] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.354839] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.354839] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.354839] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.356539] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.360665] env[61986]: INFO nova.compute.manager [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Terminating instance [ 934.365294] env[61986]: DEBUG nova.compute.manager [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.365789] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.368355] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a909faf-70c4-4b14-ae42-50f7e181d53a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.377991] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.379463] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f06090e3-efe9-4fc0-947a-e6e206d2a262 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.388967] env[61986]: DEBUG oslo_vmware.api [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 934.388967] env[61986]: value = "task-1160037" [ 934.388967] env[61986]: _type = "Task" [ 934.388967] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.403915] env[61986]: DEBUG oslo_vmware.api [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.437569] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.438021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.438232] env[61986]: DEBUG nova.network.neutron [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.446998] env[61986]: DEBUG nova.network.neutron [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Successfully updated port: 514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.541145] env[61986]: DEBUG nova.compute.manager [req-08a9d168-7a63-4258-8b2d-9e2af080919d req-a88a9b2c-3a29-4976-bd84-71c8a22cc3f3 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Received event network-vif-plugged-514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 934.541391] env[61986]: DEBUG oslo_concurrency.lockutils [req-08a9d168-7a63-4258-8b2d-9e2af080919d req-a88a9b2c-3a29-4976-bd84-71c8a22cc3f3 service nova] Acquiring lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.541745] env[61986]: DEBUG oslo_concurrency.lockutils [req-08a9d168-7a63-4258-8b2d-9e2af080919d req-a88a9b2c-3a29-4976-bd84-71c8a22cc3f3 service nova] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.542707] env[61986]: DEBUG oslo_concurrency.lockutils [req-08a9d168-7a63-4258-8b2d-9e2af080919d req-a88a9b2c-3a29-4976-bd84-71c8a22cc3f3 service nova] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.543326] env[61986]: DEBUG nova.compute.manager [req-08a9d168-7a63-4258-8b2d-9e2af080919d req-a88a9b2c-3a29-4976-bd84-71c8a22cc3f3 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] No waiting events found dispatching network-vif-plugged-514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.543555] env[61986]: WARNING nova.compute.manager [req-08a9d168-7a63-4258-8b2d-9e2af080919d req-a88a9b2c-3a29-4976-bd84-71c8a22cc3f3 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Received unexpected event network-vif-plugged-514d8051-f2f3-40b1-99fb-77d3a6b00f56 for instance with vm_state building and task_state spawning. [ 934.581544] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160032, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561907} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.587119] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] f15bf19d-b86a-4b0a-ac1a-9df8e77e9382/f15bf19d-b86a-4b0a-ac1a-9df8e77e9382.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 934.587351] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.588251] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e75ea56-f9cf-4b28-b358-425649ffa147 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.600015] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 934.600015] env[61986]: value = "task-1160038" [ 934.600015] env[61986]: _type = "Task" [ 934.600015] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.609202] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.687448] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.735841] env[61986]: DEBUG nova.objects.instance [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lazy-loading 'flavor' on Instance uuid a0fa2cbe-d8c3-462a-9744-aaa36a8d314e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.750441] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52add07b-016f-f9c1-931f-c3a1ecb6fb5a, 'name': SearchDatastore_Task, 'duration_secs': 0.019631} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.751456] env[61986]: DEBUG oslo_concurrency.lockutils [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.751456] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk. {{(pid=61986) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 934.751695] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a220a02-bb9b-4da1-a847-1d8c977fcc9d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.760012] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 934.760012] env[61986]: value = "task-1160039" [ 934.760012] env[61986]: _type = "Task" [ 934.760012] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.773071] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.838422] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160036, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.867233] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "dea128a9-aed0-40b2-ae17-c068ea8e3452" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.867516] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "dea128a9-aed0-40b2-ae17-c068ea8e3452" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.867806] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "dea128a9-aed0-40b2-ae17-c068ea8e3452-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.868081] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "dea128a9-aed0-40b2-ae17-c068ea8e3452-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.868298] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "dea128a9-aed0-40b2-ae17-c068ea8e3452-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.873728] env[61986]: INFO nova.compute.manager [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Terminating instance [ 934.876139] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "refresh_cache-dea128a9-aed0-40b2-ae17-c068ea8e3452" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.876640] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquired lock "refresh_cache-dea128a9-aed0-40b2-ae17-c068ea8e3452" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.877691] env[61986]: DEBUG nova.network.neutron [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.905388] env[61986]: DEBUG oslo_vmware.api [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160037, 'name': PowerOffVM_Task, 'duration_secs': 0.209199} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.905975] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.906313] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.906531] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b91fb7b0-b9cc-4ba8-a627-79640148ea4b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.952618] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "refresh_cache-f9aa0511-b05c-408f-ac06-c49bf8dc648c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.953714] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquired lock "refresh_cache-f9aa0511-b05c-408f-ac06-c49bf8dc648c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.953714] env[61986]: DEBUG nova.network.neutron [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.969926] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.970980] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.971351] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleting the datastore file [datastore2] 49bcac13-6dde-4a28-9131-f9a1e1d2e386 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.974906] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e340446e-8b16-491c-8b75-0bdb352ce64f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.982886] env[61986]: DEBUG oslo_vmware.api [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 934.982886] env[61986]: value = "task-1160041" [ 934.982886] env[61986]: _type = "Task" [ 934.982886] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.994971] env[61986]: DEBUG oslo_vmware.api [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.002086] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302d5fcc-2129-40ea-bf82-1f7f5f3a327e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.011535] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcdaf20-936a-42bb-ab8c-1d25603bb23d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.053031] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbd0cea-c2f7-47c0-86bf-99255327ccd6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.062179] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7ed78d-a066-41a2-8f04-201a9c9077f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.076550] env[61986]: DEBUG nova.compute.provider_tree [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.114133] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182572} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.114133] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.114133] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbf363c-cb22-424a-9b2a-666b52dad3f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.138075] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] f15bf19d-b86a-4b0a-ac1a-9df8e77e9382/f15bf19d-b86a-4b0a-ac1a-9df8e77e9382.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.141120] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06d7c31d-1a8d-4b89-ab75-b67c6750eadc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.162316] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 935.162316] env[61986]: value = "task-1160042" [ 935.162316] env[61986]: _type = "Task" [ 935.162316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.171196] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160042, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.232163] env[61986]: DEBUG nova.network.neutron [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [{"id": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "address": "fa:16:3e:5e:4f:21", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7c53e1f-f6", "ovs_interfaceid": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.270930] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160039, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.334447] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160036, 'name': Rename_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.400355] env[61986]: DEBUG nova.network.neutron [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.468714] env[61986]: DEBUG nova.network.neutron [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.494018] env[61986]: DEBUG oslo_vmware.api [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.392378} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.494771] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.495385] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.495679] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.496031] env[61986]: INFO nova.compute.manager [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Took 1.13 seconds to destroy the instance on the hypervisor. [ 935.496458] env[61986]: DEBUG oslo.service.loopingcall [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.496811] env[61986]: DEBUG nova.compute.manager [-] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.497335] env[61986]: DEBUG nova.network.neutron [-] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.551469] env[61986]: DEBUG nova.network.neutron [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.583234] env[61986]: DEBUG nova.scheduler.client.report [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.673268] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160042, 'name': ReconfigVM_Task, 'duration_secs': 0.37688} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.674875] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Reconfigured VM instance instance-0000004f to attach disk [datastore1] f15bf19d-b86a-4b0a-ac1a-9df8e77e9382/f15bf19d-b86a-4b0a-ac1a-9df8e77e9382.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.674875] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-358fe8ec-8b16-4794-a911-f5a1b344a785 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.682672] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 935.682672] env[61986]: value = "task-1160043" [ 935.682672] env[61986]: _type = "Task" [ 935.682672] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.692560] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160043, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.734193] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.745225] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7cc69aad-0187-4bef-9bdd-50acf5067117 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.877s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.771579] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.835459} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.772139] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk. [ 935.772862] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804a7c8e-76d4-4011-b818-e17d4c53df33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.803303] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.809041] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ecc3958-dfaf-4e58-8673-64315d984a52 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.825530] env[61986]: DEBUG nova.network.neutron [-] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.837746] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160036, 'name': Rename_Task, 'duration_secs': 1.192351} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.839282] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 935.839647] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 935.839647] env[61986]: value = "task-1160044" [ 935.839647] env[61986]: _type = "Task" [ 935.839647] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.839848] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2c00cdd-545c-4492-9591-f5dbbf713e0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.851393] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160044, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.853639] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 935.853639] env[61986]: value = "task-1160045" [ 935.853639] env[61986]: _type = "Task" [ 935.853639] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.864440] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.878917] env[61986]: DEBUG nova.network.neutron [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Updating instance_info_cache with network_info: [{"id": "514d8051-f2f3-40b1-99fb-77d3a6b00f56", "address": "fa:16:3e:02:3e:76", "network": {"id": "7883dcc9-3039-4100-b3df-5ffab121cc2a", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1556720724-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdbc87d45204921a43ac1555391b54a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ce1511e-1eaa-45c6-a1ef-2b714c814fa1", "external-id": "nsx-vlan-transportzone-300", "segmentation_id": 300, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap514d8051-f2", "ovs_interfaceid": "514d8051-f2f3-40b1-99fb-77d3a6b00f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.972693] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Releasing lock "refresh_cache-dea128a9-aed0-40b2-ae17-c068ea8e3452" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.973030] env[61986]: DEBUG nova.compute.manager [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 935.973246] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 935.974154] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c685e3b-0aa7-4b43-9d8d-c2f8d5bc21c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.983277] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 935.983555] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe019960-1eb6-42e9-877e-a05bc87b0be9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.990781] env[61986]: DEBUG oslo_vmware.api [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 935.990781] env[61986]: value = "task-1160046" [ 935.990781] env[61986]: _type = "Task" [ 935.990781] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.998877] env[61986]: DEBUG oslo_vmware.api [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160046, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.089956] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.180s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.093032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.309s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.094897] env[61986]: INFO nova.compute.claims [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.122132] env[61986]: INFO nova.scheduler.client.report [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted allocations for instance 6f9349b7-886a-4077-8f6d-a9800ab353fe [ 936.192809] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160043, 'name': Rename_Task, 'duration_secs': 0.190852} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.193199] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.193523] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9429941-2b3a-4199-a917-8b0c31020d1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.200308] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 936.200308] env[61986]: value = "task-1160047" [ 936.200308] env[61986]: _type = "Task" [ 936.200308] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.216070] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160047, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.329025] env[61986]: INFO nova.compute.manager [-] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Took 0.83 seconds to deallocate network for instance. [ 936.354751] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160044, 'name': ReconfigVM_Task, 'duration_secs': 0.460158} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.354995] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.363791] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d271174b-0534-4874-8f27-0c86ea24afbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.371781] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160045, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.391411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Releasing lock "refresh_cache-f9aa0511-b05c-408f-ac06-c49bf8dc648c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.392028] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Instance network_info: |[{"id": "514d8051-f2f3-40b1-99fb-77d3a6b00f56", "address": "fa:16:3e:02:3e:76", "network": {"id": "7883dcc9-3039-4100-b3df-5ffab121cc2a", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1556720724-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdbc87d45204921a43ac1555391b54a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ce1511e-1eaa-45c6-a1ef-2b714c814fa1", "external-id": "nsx-vlan-transportzone-300", "segmentation_id": 300, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap514d8051-f2", "ovs_interfaceid": "514d8051-f2f3-40b1-99fb-77d3a6b00f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 936.398214] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:3e:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ce1511e-1eaa-45c6-a1ef-2b714c814fa1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '514d8051-f2f3-40b1-99fb-77d3a6b00f56', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.406849] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Creating folder: Project (6cdbc87d45204921a43ac1555391b54a). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 936.407266] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ad46a89-c710-4916-9d17-184645f0df6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.419497] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bcfad2e-d4d5-403e-9e2c-4599d9554567 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.428179] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 936.428179] env[61986]: value = "task-1160048" [ 936.428179] env[61986]: _type = "Task" [ 936.428179] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.433861] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Created folder: Project (6cdbc87d45204921a43ac1555391b54a) in parent group-v252271. [ 936.434174] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Creating folder: Instances. Parent ref: group-v252405. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 936.437654] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3aa68592-131e-4e71-a1e4-34d85c52b620 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.440163] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160048, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.448154] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Created folder: Instances in parent group-v252405. [ 936.449260] env[61986]: DEBUG oslo.service.loopingcall [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.449508] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 936.449740] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4aff2bae-5fc9-468f-b195-4b48064b1dfb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.470458] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.470458] env[61986]: value = "task-1160051" [ 936.470458] env[61986]: _type = "Task" [ 936.470458] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.478933] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160051, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.501849] env[61986]: DEBUG oslo_vmware.api [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160046, 'name': PowerOffVM_Task, 'duration_secs': 0.145958} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.501849] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 936.501849] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 936.502141] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8dd48b4-e11d-4135-b776-a272083dae8e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.534477] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 936.534706] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 936.534920] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Deleting the datastore file [datastore1] dea128a9-aed0-40b2-ae17-c068ea8e3452 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.535225] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c6127ae-447e-459d-8f9e-ed5844d0c2ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.542476] env[61986]: DEBUG oslo_vmware.api [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for the task: (returnval){ [ 936.542476] env[61986]: value = "task-1160053" [ 936.542476] env[61986]: _type = "Task" [ 936.542476] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.550664] env[61986]: DEBUG oslo_vmware.api [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.634155] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c5c96345-6ef9-4c46-8aae-44f65a35e84b tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "6f9349b7-886a-4077-8f6d-a9800ab353fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.853s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.715036] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160047, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.836596] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.864675] env[61986]: DEBUG oslo_vmware.api [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160045, 'name': PowerOnVM_Task, 'duration_secs': 0.589132} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.864956] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 936.865362] env[61986]: INFO nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Took 9.78 seconds to spawn the instance on the hypervisor. [ 936.865613] env[61986]: DEBUG nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.866394] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ef6280-7772-4e9a-91fb-4c2ab5da8366 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.937507] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160048, 'name': ReconfigVM_Task, 'duration_secs': 0.260476} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.937794] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.938246] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70191247-eeaa-4a6d-b940-c659c0f91905 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.943711] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 936.943711] env[61986]: value = "task-1160054" [ 936.943711] env[61986]: _type = "Task" [ 936.943711] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.954094] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.980908] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160051, 'name': CreateVM_Task, 'duration_secs': 0.488297} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.981087] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 936.981797] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.981956] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.982304] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.982573] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-832b4771-5266-4961-a8f1-ed5bc542ce88 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.987463] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 936.987463] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529fbad9-5182-518d-686c-8e974a73e2b4" [ 936.987463] env[61986]: _type = "Task" [ 936.987463] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.995164] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529fbad9-5182-518d-686c-8e974a73e2b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.052385] env[61986]: DEBUG oslo_vmware.api [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Task: {'id': task-1160053, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146734} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.054347] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.054810] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 937.055017] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 937.055198] env[61986]: INFO nova.compute.manager [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Took 1.08 seconds to destroy the instance on the hypervisor. [ 937.055441] env[61986]: DEBUG oslo.service.loopingcall [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.060159] env[61986]: DEBUG nova.compute.manager [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Received event network-changed-514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 937.060159] env[61986]: DEBUG nova.compute.manager [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Refreshing instance network info cache due to event network-changed-514d8051-f2f3-40b1-99fb-77d3a6b00f56. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 937.060159] env[61986]: DEBUG oslo_concurrency.lockutils [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] Acquiring lock "refresh_cache-f9aa0511-b05c-408f-ac06-c49bf8dc648c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.060329] env[61986]: DEBUG oslo_concurrency.lockutils [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] Acquired lock "refresh_cache-f9aa0511-b05c-408f-ac06-c49bf8dc648c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.060469] env[61986]: DEBUG nova.network.neutron [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Refreshing network info cache for port 514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.061784] env[61986]: DEBUG nova.compute.manager [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.061884] env[61986]: DEBUG nova.network.neutron [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 937.077822] env[61986]: DEBUG nova.network.neutron [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 937.213741] env[61986]: DEBUG oslo_vmware.api [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160047, 'name': PowerOnVM_Task, 'duration_secs': 0.51425} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.214019] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 937.214233] env[61986]: INFO nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Took 9.10 seconds to spawn the instance on the hypervisor. [ 937.215990] env[61986]: DEBUG nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.215990] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a24516c-8307-424a-a783-d31294d1121d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.255760] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78cf14d-c3cc-488d-84a0-4ab88c62c09a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.280423] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 937.387688] env[61986]: INFO nova.compute.manager [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Took 30.20 seconds to build instance. [ 937.458570] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160054, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.494821] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e39be2-75cd-4418-b671-7ab125d49935 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.504096] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529fbad9-5182-518d-686c-8e974a73e2b4, 'name': SearchDatastore_Task, 'duration_secs': 0.041432} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.506245] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.506502] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.506741] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.506893] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.507106] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.507473] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e1e7ca8-5ee2-4520-bccf-6f09d2a53a23 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.510097] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278cea73-12be-4018-bbd8-4d2950818375 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.550832] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fbcb46-9d35-4fbe-a8d0-358ce9a8d360 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.553811] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.553996] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 937.554766] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-194743ae-2eaf-4997-9839-e9df94e6b5db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.562147] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.562393] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.562605] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.562795] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.563031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.568443] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602cb12e-23d8-4fc3-9853-5be4d443ab83 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.572432] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 937.572432] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f8d65c-c8d4-d3c2-a00f-01852dd90bb1" [ 937.572432] env[61986]: _type = "Task" [ 937.572432] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.572970] env[61986]: INFO nova.compute.manager [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Terminating instance [ 937.587579] env[61986]: DEBUG nova.compute.manager [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.587805] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 937.588439] env[61986]: DEBUG nova.network.neutron [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.589829] env[61986]: DEBUG nova.compute.provider_tree [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.594229] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295a8dc3-2cc2-445a-ad8e-398b513fcfa6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.601777] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f8d65c-c8d4-d3c2-a00f-01852dd90bb1, 'name': SearchDatastore_Task, 'duration_secs': 0.020425} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.605327] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 937.608249] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4f170b9-ae8a-4f1f-bf65-8af4c0996230 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.612061] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d3accb9-dca5-4d1d-a32d-d940dece3f3d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.628470] env[61986]: DEBUG oslo_vmware.api [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 937.628470] env[61986]: value = "task-1160055" [ 937.628470] env[61986]: _type = "Task" [ 937.628470] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.630525] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 937.630525] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e64a59-3bfa-62ab-7c2f-870055dfb9ce" [ 937.630525] env[61986]: _type = "Task" [ 937.630525] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.649732] env[61986]: DEBUG oslo_vmware.api [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.652598] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e64a59-3bfa-62ab-7c2f-870055dfb9ce, 'name': SearchDatastore_Task, 'duration_secs': 0.013811} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.652749] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.652997] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] f9aa0511-b05c-408f-ac06-c49bf8dc648c/f9aa0511-b05c-408f-ac06-c49bf8dc648c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 937.653292] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f9b9552-db7b-4ba6-8002-b82705405020 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.663172] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 937.663172] env[61986]: value = "task-1160056" [ 937.663172] env[61986]: _type = "Task" [ 937.663172] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.672106] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160056, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.740392] env[61986]: INFO nova.compute.manager [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Took 29.26 seconds to build instance. [ 937.788432] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 937.789153] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6154697c-851d-4ded-8182-119a78c36221 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.796618] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 937.796618] env[61986]: value = "task-1160057" [ 937.796618] env[61986]: _type = "Task" [ 937.796618] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.809368] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.889999] env[61986]: DEBUG oslo_concurrency.lockutils [None req-98a95c34-21b9-45e4-9aa5-f15d79f469c5 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.597s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.904612] env[61986]: DEBUG nova.network.neutron [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Updated VIF entry in instance network info cache for port 514d8051-f2f3-40b1-99fb-77d3a6b00f56. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.905177] env[61986]: DEBUG nova.network.neutron [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Updating instance_info_cache with network_info: [{"id": "514d8051-f2f3-40b1-99fb-77d3a6b00f56", "address": "fa:16:3e:02:3e:76", "network": {"id": "7883dcc9-3039-4100-b3df-5ffab121cc2a", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1556720724-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cdbc87d45204921a43ac1555391b54a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ce1511e-1eaa-45c6-a1ef-2b714c814fa1", "external-id": "nsx-vlan-transportzone-300", "segmentation_id": 300, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap514d8051-f2", "ovs_interfaceid": "514d8051-f2f3-40b1-99fb-77d3a6b00f56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.955690] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160054, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.096598] env[61986]: DEBUG nova.scheduler.client.report [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.101115] env[61986]: INFO nova.compute.manager [-] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Took 1.04 seconds to deallocate network for instance. [ 938.145371] env[61986]: DEBUG oslo_vmware.api [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160055, 'name': PowerOffVM_Task, 'duration_secs': 0.4918} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.145668] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.145849] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 938.146159] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd774551-c74e-4943-92a9-87345bda1e13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.178804] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.178971] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.179301] env[61986]: DEBUG nova.compute.manager [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.179631] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160056, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.180459] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc2e1d8-a941-494a-9086-48c0e8a0a828 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.188055] env[61986]: DEBUG nova.compute.manager [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61986) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 938.188722] env[61986]: DEBUG nova.objects.instance [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'flavor' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.213353] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 938.213617] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 938.213767] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Deleting the datastore file [datastore1] a0fa2cbe-d8c3-462a-9744-aaa36a8d314e {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.214058] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4495e96-7ad2-43a2-9459-a61234e169c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.221032] env[61986]: DEBUG oslo_vmware.api [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for the task: (returnval){ [ 938.221032] env[61986]: value = "task-1160059" [ 938.221032] env[61986]: _type = "Task" [ 938.221032] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.229850] env[61986]: DEBUG oslo_vmware.api [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.243326] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56608030-62a3-4bc3-9d42-6921ffb367bc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.633s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.308308] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160057, 'name': PowerOffVM_Task, 'duration_secs': 0.315233} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.308436] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.308606] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 938.409528] env[61986]: DEBUG oslo_concurrency.lockutils [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] Releasing lock "refresh_cache-f9aa0511-b05c-408f-ac06-c49bf8dc648c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.409879] env[61986]: DEBUG nova.compute.manager [req-964cc165-5fea-4c07-a102-d82ab21ee9de req-0d004163-6804-4120-94d2-1bdc14392548 service nova] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Received event network-vif-deleted-495a4f17-b010-4427-bf0f-e72582cd1dc1 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 938.457133] env[61986]: DEBUG oslo_vmware.api [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160054, 'name': PowerOnVM_Task, 'duration_secs': 1.417744} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.457133] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.460484] env[61986]: DEBUG nova.compute.manager [None req-347864d5-0d82-43f4-a864-05a90a432667 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.461959] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b8e495-2db0-43c7-b791-8d53ab1ab653 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.603237] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.603834] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.606588] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.269s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.606815] env[61986]: DEBUG nova.objects.instance [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lazy-loading 'resources' on Instance uuid 3512cd72-6666-4810-828e-50230956c4b1 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.610972] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.673441] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.673675] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.674224] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.674224] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.674224] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.676160] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160056, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.720902} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.676376] env[61986]: INFO nova.compute.manager [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Terminating instance [ 938.677747] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] f9aa0511-b05c-408f-ac06-c49bf8dc648c/f9aa0511-b05c-408f-ac06-c49bf8dc648c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 938.677960] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.678547] env[61986]: DEBUG nova.compute.manager [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.678729] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.678957] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2f96444-1924-4744-a502-bf0cd4e9ed86 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.681789] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c32c56-d1e9-4f22-88da-ab4228f21b97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.688902] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.690036] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad208315-9d01-4c86-81bb-c58b0bb7c4f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.691580] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 938.691580] env[61986]: value = "task-1160060" [ 938.691580] env[61986]: _type = "Task" [ 938.691580] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.698693] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.699045] env[61986]: DEBUG oslo_vmware.api [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 938.699045] env[61986]: value = "task-1160061" [ 938.699045] env[61986]: _type = "Task" [ 938.699045] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.699341] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce0f3ba8-01ab-457e-bd42-366d35b17e5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.708497] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160060, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.710308] env[61986]: DEBUG oslo_vmware.api [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 938.710308] env[61986]: value = "task-1160062" [ 938.710308] env[61986]: _type = "Task" [ 938.710308] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.716889] env[61986]: DEBUG oslo_vmware.api [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.721839] env[61986]: DEBUG oslo_vmware.api [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.729872] env[61986]: DEBUG oslo_vmware.api [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Task: {'id': task-1160059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.481839} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.730201] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.730399] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 938.730582] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 938.730765] env[61986]: INFO nova.compute.manager [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 938.731063] env[61986]: DEBUG oslo.service.loopingcall [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.731266] env[61986]: DEBUG nova.compute.manager [-] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 938.731417] env[61986]: DEBUG nova.network.neutron [-] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 938.816355] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.816700] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.816887] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.817779] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.817779] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.817779] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.821160] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.821160] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.821160] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.821160] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.821160] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.827114] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8db1a090-dc65-4955-8383-1dea8952c593 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.847123] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 938.847123] env[61986]: value = "task-1160063" [ 938.847123] env[61986]: _type = "Task" [ 938.847123] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.861649] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160063, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.974457] env[61986]: INFO nova.compute.manager [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Rebuilding instance [ 939.015925] env[61986]: DEBUG nova.compute.manager [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.017047] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b93a460-be91-4cc3-b794-35a5bd9ae71e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.116230] env[61986]: DEBUG nova.compute.utils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.121260] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.121260] env[61986]: DEBUG nova.network.neutron [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 939.162994] env[61986]: DEBUG nova.policy [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 939.212995] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160060, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072796} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.218627] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.223534] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce10c2a-8ee0-4749-b9af-2e4b13c51275 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.229922] env[61986]: DEBUG oslo_vmware.api [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160061, 'name': PowerOffVM_Task, 'duration_secs': 0.237386} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.233222] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.233222] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.234231] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00feafdd-ef68-4821-8833-5219bc912aa8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.265919] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] f9aa0511-b05c-408f-ac06-c49bf8dc648c/f9aa0511-b05c-408f-ac06-c49bf8dc648c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.266941] env[61986]: DEBUG oslo_vmware.api [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160062, 'name': PowerOffVM_Task, 'duration_secs': 0.248044} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.270040] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8303771e-c821-43ed-81e5-64a7058ff3ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.289225] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.289225] env[61986]: DEBUG nova.compute.manager [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.290845] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da9ad17-33c6-4a73-8fb1-5dce9f4d17e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.304875] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 939.304875] env[61986]: value = "task-1160065" [ 939.304875] env[61986]: _type = "Task" [ 939.304875] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.318479] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160065, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.339413] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.339413] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.339413] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleting the datastore file [datastore1] eaa61f81-2f9b-4d1c-bab8-4363fb71e936 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.339759] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-daeea91a-81c2-4800-85bf-97cfcbd17102 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.348886] env[61986]: DEBUG oslo_vmware.api [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 939.348886] env[61986]: value = "task-1160066" [ 939.348886] env[61986]: _type = "Task" [ 939.348886] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.369894] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160063, 'name': ReconfigVM_Task, 'duration_secs': 0.196165} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.374102] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 939.378951] env[61986]: DEBUG oslo_vmware.api [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160066, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.532555] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.533597] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d6caa98-4b84-4086-824e-beb02cd9948c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.543983] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.544282] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.546097] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 939.546097] env[61986]: value = "task-1160067" [ 939.546097] env[61986]: _type = "Task" [ 939.546097] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.560160] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160067, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.611396] env[61986]: DEBUG nova.network.neutron [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Successfully created port: 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.623280] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.735692] env[61986]: DEBUG nova.compute.manager [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Received event network-changed-8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 939.735950] env[61986]: DEBUG nova.compute.manager [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Refreshing instance network info cache due to event network-changed-8157a838-3c36-4936-a187-460e14d7f129. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 939.736317] env[61986]: DEBUG oslo_concurrency.lockutils [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] Acquiring lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.736358] env[61986]: DEBUG oslo_concurrency.lockutils [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] Acquired lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.736567] env[61986]: DEBUG nova.network.neutron [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Refreshing network info cache for port 8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 939.739666] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f3718b-2491-4a03-85af-7b9a9d4f72d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.750604] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54068f09-85e8-4c8a-b2c9-bb800ae33756 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.783705] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6745690-1cbc-4309-863f-f98ac919c404 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.791523] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03258003-fce3-46ea-8111-1adbd7a39793 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.808195] env[61986]: DEBUG nova.compute.provider_tree [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.809539] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f2f3d7c8-e108-4ba9-a613-f4b643a2e27d tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.631s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.819124] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160065, 'name': ReconfigVM_Task, 'duration_secs': 0.426849} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.819977] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Reconfigured VM instance instance-00000050 to attach disk [datastore1] f9aa0511-b05c-408f-ac06-c49bf8dc648c/f9aa0511-b05c-408f-ac06-c49bf8dc648c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.820626] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae86bb6f-27a0-41e8-8423-732157876125 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.827227] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 939.827227] env[61986]: value = "task-1160068" [ 939.827227] env[61986]: _type = "Task" [ 939.827227] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.835950] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160068, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.867174] env[61986]: DEBUG oslo_vmware.api [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160066, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242367} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.867567] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.867711] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.867838] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.868041] env[61986]: INFO nova.compute.manager [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Took 1.19 seconds to destroy the instance on the hypervisor. [ 939.868310] env[61986]: DEBUG oslo.service.loopingcall [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.868636] env[61986]: DEBUG nova.compute.manager [-] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.868636] env[61986]: DEBUG nova.network.neutron [-] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.880592] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.880838] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.880999] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.881202] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.881408] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.881501] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.881708] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.881874] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.882063] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.882244] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.882906] env[61986]: DEBUG nova.virt.hardware [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.887929] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Reconfiguring VM instance instance-00000046 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 939.888242] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5806056e-77f6-48c4-9dc2-853ab00267d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.906623] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 939.906623] env[61986]: value = "task-1160069" [ 939.906623] env[61986]: _type = "Task" [ 939.906623] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.915829] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160069, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.050930] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.067213] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160067, 'name': PowerOffVM_Task, 'duration_secs': 0.279006} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.067213] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 940.067213] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 940.067213] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d112947-537b-4362-a388-f37f71e40706 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.074784] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 940.078017] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-192e45f7-8853-46c3-b179-be5257212ddd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.143505] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 940.143505] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 940.143678] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleting the datastore file [datastore2] aa12919e-3891-4da9-a280-4155137864e1 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.143845] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf9f9b84-c4b6-4bc0-af75-8ee57fe10549 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.152259] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 940.152259] env[61986]: value = "task-1160071" [ 940.152259] env[61986]: _type = "Task" [ 940.152259] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.160765] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.242358] env[61986]: DEBUG nova.compute.manager [req-8efd638f-69f8-44e3-9e46-e468184a6509 req-43c9c920-e792-4c9d-ac03-5017c3334ca2 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Received event network-vif-deleted-3cf0a017-8fc3-4649-bb5a-922ad76c9140 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.242358] env[61986]: INFO nova.compute.manager [req-8efd638f-69f8-44e3-9e46-e468184a6509 req-43c9c920-e792-4c9d-ac03-5017c3334ca2 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Neutron deleted interface 3cf0a017-8fc3-4649-bb5a-922ad76c9140; detaching it from the instance and deleting it from the info cache [ 940.242358] env[61986]: DEBUG nova.network.neutron [req-8efd638f-69f8-44e3-9e46-e468184a6509 req-43c9c920-e792-4c9d-ac03-5017c3334ca2 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.317019] env[61986]: DEBUG nova.scheduler.client.report [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.341450] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160068, 'name': Rename_Task, 'duration_secs': 0.267445} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.341740] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 940.342430] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6599d14-dcb2-4534-b459-9086f1a92c54 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.348888] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 940.348888] env[61986]: value = "task-1160072" [ 940.348888] env[61986]: _type = "Task" [ 940.348888] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.358060] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160072, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.432617] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160069, 'name': ReconfigVM_Task, 'duration_secs': 0.276712} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.433464] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Reconfigured VM instance instance-00000046 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 940.433943] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e3e35d-0dee-4e5c-b275-3fc6b1510bbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.467439] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee/f42e92ef-8a90-473c-8662-57cbb9e3f4ee.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.471248] env[61986]: DEBUG nova.network.neutron [-] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.471698] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c999f39-95f7-4564-b0bd-e4bad6550f9b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.490711] env[61986]: INFO nova.compute.manager [-] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Took 1.76 seconds to deallocate network for instance. [ 940.501019] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 940.501019] env[61986]: value = "task-1160073" [ 940.501019] env[61986]: _type = "Task" [ 940.501019] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.515992] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160073, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.598787] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.634632] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.670820] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147337} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.673932] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.674267] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.674485] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.674717] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.674913] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.675133] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.675396] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.675606] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.675830] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.676060] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.676439] env[61986]: DEBUG nova.virt.hardware [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.676815] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.677060] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.677464] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.680758] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216eb0bf-1f33-46e5-bb69-259b9130bcea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.690699] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47993013-b630-4d15-bb3d-f61f0b0892c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.747972] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b06457c-5e0a-43fc-bc42-79e74275ccd5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.751154] env[61986]: DEBUG nova.network.neutron [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updated VIF entry in instance network info cache for port 8157a838-3c36-4936-a187-460e14d7f129. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.751595] env[61986]: DEBUG nova.network.neutron [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updating instance_info_cache with network_info: [{"id": "8157a838-3c36-4936-a187-460e14d7f129", "address": "fa:16:3e:c6:10:13", "network": {"id": "33c1c5c8-89b8-4be8-a7e6-6411f9c40a9c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1837194892-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fe9c887c0d3485f8e6f01785b2a4743", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b89fd3b-0470-40c9-bb5b-d52c76c030e4", "external-id": "nsx-vlan-transportzone-276", "segmentation_id": 276, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8157a838-3c", "ovs_interfaceid": "8157a838-3c36-4936-a187-460e14d7f129", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.761192] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8246ff-200d-4eda-b05e-96a3e0358c7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.801945] env[61986]: DEBUG nova.compute.manager [req-8efd638f-69f8-44e3-9e46-e468184a6509 req-43c9c920-e792-4c9d-ac03-5017c3334ca2 service nova] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Detach interface failed, port_id=3cf0a017-8fc3-4649-bb5a-922ad76c9140, reason: Instance a0fa2cbe-d8c3-462a-9744-aaa36a8d314e could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 940.823115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.216s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.825041] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.283s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.825306] env[61986]: DEBUG nova.objects.instance [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lazy-loading 'resources' on Instance uuid b833adc9-2967-4b0b-81c2-0b8deac20f69 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.843318] env[61986]: INFO nova.scheduler.client.report [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted allocations for instance 3512cd72-6666-4810-828e-50230956c4b1 [ 940.862216] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160072, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.912123] env[61986]: DEBUG nova.network.neutron [-] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.002538] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.012326] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160073, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.254683] env[61986]: DEBUG oslo_concurrency.lockutils [req-65bfe25f-392a-4a1d-b8ff-a0dfad384b7e req-685d05cf-87ec-4f29-91f9-d86fd6e58ab2 service nova] Releasing lock "refresh_cache-5170aa51-3307-42b1-b0dd-645dd4036e5b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.264469] env[61986]: DEBUG nova.network.neutron [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Successfully updated port: 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.281426] env[61986]: INFO nova.compute.manager [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Unrescuing [ 941.281426] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.281426] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquired lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.281426] env[61986]: DEBUG nova.network.neutron [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.357948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b3bf16f0-3a39-49df-bfb9-abbbbe9f02d8 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "3512cd72-6666-4810-828e-50230956c4b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.037s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.376317] env[61986]: DEBUG oslo_vmware.api [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160072, 'name': PowerOnVM_Task, 'duration_secs': 0.720285} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.376745] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 941.377250] env[61986]: INFO nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Took 8.18 seconds to spawn the instance on the hypervisor. [ 941.377741] env[61986]: DEBUG nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.380279] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768784be-4057-45a3-b8ac-adcc7dd637b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.414371] env[61986]: INFO nova.compute.manager [-] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Took 1.55 seconds to deallocate network for instance. [ 941.514303] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160073, 'name': ReconfigVM_Task, 'duration_secs': 0.55203} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.514689] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Reconfigured VM instance instance-00000046 to attach disk [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee/f42e92ef-8a90-473c-8662-57cbb9e3f4ee.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.515727] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 50 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 941.720043] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.720043] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.720618] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.720618] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.720618] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.720763] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.721038] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.721038] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.721218] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.721374] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.721559] env[61986]: DEBUG nova.virt.hardware [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.722468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36de1246-5cb2-473a-be85-eeea4e62dc2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.731478] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec4529b-6466-4541-b945-acafc57d9c86 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.747939] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:61:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '038ca95f-90ee-4bd7-afb9-b514378d344e', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.755271] env[61986]: DEBUG oslo.service.loopingcall [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.755615] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 941.755892] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9310554-c54f-4248-af91-2eb34d7388eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.774132] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.774278] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.774422] env[61986]: DEBUG nova.network.neutron [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.776656] env[61986]: DEBUG nova.compute.manager [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Received event network-vif-deleted-86c32f0f-e39d-4b76-a6b6-a0d93b2e9e67 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.776866] env[61986]: DEBUG nova.compute.manager [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-vif-plugged-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.777066] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.777303] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.777478] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.777648] env[61986]: DEBUG nova.compute.manager [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] No waiting events found dispatching network-vif-plugged-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.777819] env[61986]: WARNING nova.compute.manager [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received unexpected event network-vif-plugged-51b89bbc-3efd-4e51-b21b-bfce45546eae for instance with vm_state building and task_state spawning. [ 941.777984] env[61986]: DEBUG nova.compute.manager [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.778159] env[61986]: DEBUG nova.compute.manager [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing instance network info cache due to event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 941.778333] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.779396] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053fe672-fd63-42be-97d8-8fb66991b763 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.787740] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.787740] env[61986]: value = "task-1160074" [ 941.787740] env[61986]: _type = "Task" [ 941.787740] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.794264] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0312a5de-b95a-447c-af1d-2e6862e378a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.802652] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160074, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.832163] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97138c2-dfe8-4747-876b-11b6032356a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.839219] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b773b95e-3372-4cd2-a5d8-3296f3a7005f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.853478] env[61986]: DEBUG nova.compute.provider_tree [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.914894] env[61986]: INFO nova.compute.manager [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Took 33.05 seconds to build instance. [ 941.918840] env[61986]: DEBUG nova.compute.manager [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Stashing vm_state: stopped {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 941.923701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.018473] env[61986]: DEBUG nova.network.neutron [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updating instance_info_cache with network_info: [{"id": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "address": "fa:16:3e:41:aa:c3", "network": {"id": "91655731-d34e-40b1-8ed6-d346b25296b8", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-583187102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0250d1bd7f4147a19dcb32797b5d794b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4411312-86", "ovs_interfaceid": "c4411312-86a5-48dd-a108-5c5bc29d5f3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.026905] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8a41de-b8b8-44ab-b4cf-26d26b6ad4b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.057442] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66c1240-49cb-4c8e-b12e-3545a2e7b4f0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.078175] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 67 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 942.271330] env[61986]: DEBUG nova.compute.manager [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Received event network-changed-a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 942.271605] env[61986]: DEBUG nova.compute.manager [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Refreshing instance network info cache due to event network-changed-a9804290-f49e-46f6-90c6-3ce3e529d094. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 942.271821] env[61986]: DEBUG oslo_concurrency.lockutils [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] Acquiring lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.271986] env[61986]: DEBUG oslo_concurrency.lockutils [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] Acquired lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.272170] env[61986]: DEBUG nova.network.neutron [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Refreshing network info cache for port a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.299788] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160074, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.319382] env[61986]: DEBUG nova.network.neutron [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 942.357104] env[61986]: DEBUG nova.scheduler.client.report [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.413449] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.418993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3ad72d0a-f200-429a-b144-3e6e04ab792e tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.871s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.419336] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.006s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.419603] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.419831] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.420042] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.422422] env[61986]: INFO nova.compute.manager [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Terminating instance [ 942.427461] env[61986]: DEBUG nova.compute.manager [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.427918] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 942.428649] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5ef577-d7ee-426c-af7c-0129613be5fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.436623] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 942.436623] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c897e3a-3f98-4cbe-8737-67d64d39a900 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.444114] env[61986]: DEBUG oslo_vmware.api [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 942.444114] env[61986]: value = "task-1160075" [ 942.444114] env[61986]: _type = "Task" [ 942.444114] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.445037] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.453912] env[61986]: DEBUG oslo_vmware.api [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.485473] env[61986]: DEBUG nova.network.neutron [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.522132] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Releasing lock "refresh_cache-93c4f104-1812-4bb7-bfa7-cbf70a19ff51" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.522806] env[61986]: DEBUG nova.objects.instance [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lazy-loading 'flavor' on Instance uuid 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.621276] env[61986]: DEBUG nova.network.neutron [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Port e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 942.805493] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160074, 'name': CreateVM_Task, 'duration_secs': 0.528104} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.805493] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 942.805493] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.805493] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.805493] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.805493] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91931d58-cb54-4943-b2bc-0dcae1af8044 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.808765] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 942.808765] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b240e6-90b4-79be-5251-10b3f94c6224" [ 942.808765] env[61986]: _type = "Task" [ 942.808765] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.822586] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b240e6-90b4-79be-5251-10b3f94c6224, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.868291] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.043s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.871583] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.609s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.874194] env[61986]: DEBUG nova.objects.instance [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lazy-loading 'resources' on Instance uuid 076c8e91-af70-489f-89d5-f598166fafc6 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.895121] env[61986]: INFO nova.scheduler.client.report [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Deleted allocations for instance b833adc9-2967-4b0b-81c2-0b8deac20f69 [ 942.955110] env[61986]: DEBUG oslo_vmware.api [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160075, 'name': PowerOffVM_Task, 'duration_secs': 0.277045} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.955394] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.955568] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 942.955830] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de5b7e81-b24d-4794-a26b-e671dd4a203f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.988305] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.988761] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Instance network_info: |[{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.989259] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.989551] env[61986]: DEBUG nova.network.neutron [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.990933] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:b1:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e0cfc48-d93b-4477-8082-69a2f7aa7701', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51b89bbc-3efd-4e51-b21b-bfce45546eae', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.998523] env[61986]: DEBUG oslo.service.loopingcall [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.999490] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.999720] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f750cef-88c3-4081-a9fb-21d5e0dfb531 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.022313] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.022313] env[61986]: value = "task-1160077" [ 943.022313] env[61986]: _type = "Task" [ 943.022313] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.029386] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cb0f42-a21e-4daf-a04c-6e62a6787bb7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.035818] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160077, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.037508] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 943.037713] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 943.037894] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Deleting the datastore file [datastore1] f9aa0511-b05c-408f-ac06-c49bf8dc648c {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.038164] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b5befdd-253f-4725-a1b3-b67b8d5481c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.055935] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 943.056967] env[61986]: DEBUG nova.network.neutron [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Updated VIF entry in instance network info cache for port a9804290-f49e-46f6-90c6-3ce3e529d094. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.057310] env[61986]: DEBUG nova.network.neutron [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Updating instance_info_cache with network_info: [{"id": "a9804290-f49e-46f6-90c6-3ce3e529d094", "address": "fa:16:3e:e5:e6:a1", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9804290-f4", "ovs_interfaceid": "a9804290-f49e-46f6-90c6-3ce3e529d094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.058866] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02fffb48-00b0-436d-a5f3-a3ae81ee23f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.064616] env[61986]: DEBUG oslo_vmware.api [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for the task: (returnval){ [ 943.064616] env[61986]: value = "task-1160078" [ 943.064616] env[61986]: _type = "Task" [ 943.064616] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.069460] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 943.069460] env[61986]: value = "task-1160079" [ 943.069460] env[61986]: _type = "Task" [ 943.069460] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.080496] env[61986]: DEBUG oslo_vmware.api [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.087355] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.324730] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b240e6-90b4-79be-5251-10b3f94c6224, 'name': SearchDatastore_Task, 'duration_secs': 0.022756} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.325504] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.326027] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.326454] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.326772] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.327130] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.327603] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-500b27bc-dab3-4d4c-b4b0-c3bba9a7b733 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.343013] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.343013] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.345119] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-625f74ee-e5d4-43d3-911c-5903e331b21e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.351639] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 943.351639] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b25a6d-86e9-eec3-d1e1-d12f8fbb545d" [ 943.351639] env[61986]: _type = "Task" [ 943.351639] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.362073] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b25a6d-86e9-eec3-d1e1-d12f8fbb545d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.404296] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bbb909da-c683-41ad-b25d-8eb39fe800ca tempest-ServersNegativeTestMultiTenantJSON-1320728404 tempest-ServersNegativeTestMultiTenantJSON-1320728404-project-member] Lock "b833adc9-2967-4b0b-81c2-0b8deac20f69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.222s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.536359] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160077, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.561573] env[61986]: DEBUG oslo_concurrency.lockutils [req-a449b1e2-c991-448e-b8fb-bbd88f4712b5 req-3615a9a3-7472-40e0-afd1-e8f0a3b65e64 service nova] Releasing lock "refresh_cache-f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.574552] env[61986]: DEBUG oslo_vmware.api [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Task: {'id': task-1160078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216122} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.578322] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.578614] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 943.578891] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 943.579399] env[61986]: INFO nova.compute.manager [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 943.579772] env[61986]: DEBUG oslo.service.loopingcall [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.582461] env[61986]: DEBUG nova.compute.manager [-] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 943.582647] env[61986]: DEBUG nova.network.neutron [-] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 943.589081] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160079, 'name': PowerOffVM_Task, 'duration_secs': 0.281224} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.589379] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 943.595017] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 943.597689] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d9c1c96-ef23-4928-a812-a0c6194103ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.619014] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 943.619014] env[61986]: value = "task-1160080" [ 943.619014] env[61986]: _type = "Task" [ 943.619014] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.628818] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.642954] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.643324] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.643574] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.782335] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d477fb6f-21c1-467a-8a9a-f4638b14a883 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.791097] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99c7c84-4cc5-4587-b72e-63f108556a1e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.825676] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2d09ce-59be-446a-9025-0b5e98bbceca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.833448] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e41f43f-7add-424d-b45f-2cfb74f84a13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.849015] env[61986]: DEBUG nova.compute.provider_tree [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.859941] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b25a6d-86e9-eec3-d1e1-d12f8fbb545d, 'name': SearchDatastore_Task, 'duration_secs': 0.009783} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.861379] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-688d32af-ffa6-4836-921a-ab60314e7db4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.867170] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 943.867170] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52367b3e-3ccd-84f0-2d9e-61bbd6881174" [ 943.867170] env[61986]: _type = "Task" [ 943.867170] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.875018] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52367b3e-3ccd-84f0-2d9e-61bbd6881174, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.888335] env[61986]: DEBUG nova.network.neutron [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated VIF entry in instance network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.888696] env[61986]: DEBUG nova.network.neutron [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.034301] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160077, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.129319] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.352253] env[61986]: DEBUG nova.scheduler.client.report [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.379937] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52367b3e-3ccd-84f0-2d9e-61bbd6881174, 'name': SearchDatastore_Task, 'duration_secs': 0.009333} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.383042] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.383042] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.383042] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f379cd18-2cb6-4961-b47e-1b5e58e3514a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.389077] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 944.389077] env[61986]: value = "task-1160081" [ 944.389077] env[61986]: _type = "Task" [ 944.389077] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.393857] env[61986]: DEBUG oslo_concurrency.lockutils [req-0bdda074-089c-4c1f-977d-f57df1168234 req-df7b170c-7a13-4de1-a130-8595fc2854b5 service nova] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.400452] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.449737] env[61986]: DEBUG nova.network.neutron [-] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.535377] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160077, 'name': CreateVM_Task, 'duration_secs': 1.452612} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.535588] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 944.537067] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.537290] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.537662] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.539242] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a728284e-3f54-4067-9968-6d76a011da05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.543066] env[61986]: DEBUG nova.compute.manager [req-39c65ad3-6462-41dd-9adc-ee80f5f1fc7f req-9935cff4-0d0d-4035-af6b-a60a02a6c6ea service nova] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Received event network-vif-deleted-514d8051-f2f3-40b1-99fb-77d3a6b00f56 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 944.547418] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 944.547418] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52089d0e-2f60-3de1-75d9-82c8349fc169" [ 944.547418] env[61986]: _type = "Task" [ 944.547418] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.557267] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52089d0e-2f60-3de1-75d9-82c8349fc169, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.608380] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52714534-8578-5599-14e2-ae546d642966/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 944.610864] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60418dc2-abec-4f7d-a8b8-93adbad73f66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.618074] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52714534-8578-5599-14e2-ae546d642966/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 944.618399] env[61986]: ERROR oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52714534-8578-5599-14e2-ae546d642966/disk-0.vmdk due to incomplete transfer. [ 944.618702] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c4755edc-b137-41e2-a77d-9ab65fd75c49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.631020] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.635024] env[61986]: DEBUG oslo_vmware.rw_handles [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52714534-8578-5599-14e2-ae546d642966/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 944.635024] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Uploaded image 600544b5-22c3-4141-8687-1b96be17aca7 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 944.636289] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 944.636705] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b6127b15-8323-4750-8b13-cfde5d35be1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.643875] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 944.643875] env[61986]: value = "task-1160082" [ 944.643875] env[61986]: _type = "Task" [ 944.643875] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.657587] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160082, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.740760] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.741012] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.741249] env[61986]: DEBUG nova.network.neutron [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.858680] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.863566] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.121s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.863940] env[61986]: DEBUG nova.objects.instance [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lazy-loading 'resources' on Instance uuid 52eda388-05f4-416c-ac9d-bd1c1e31ba9c {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.901180] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160081, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.903429] env[61986]: INFO nova.scheduler.client.report [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleted allocations for instance 076c8e91-af70-489f-89d5-f598166fafc6 [ 944.951657] env[61986]: INFO nova.compute.manager [-] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Took 1.37 seconds to deallocate network for instance. [ 945.058703] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52089d0e-2f60-3de1-75d9-82c8349fc169, 'name': SearchDatastore_Task, 'duration_secs': 0.011859} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.059237] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.059799] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.060253] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.060379] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.060631] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.062744] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2bbdde1-e82b-4f06-977d-2627a9317669 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.071668] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.071820] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 945.072532] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7e81a82-cdfb-496c-86cb-56960e299f32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.078226] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 945.078226] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52dcd6f6-1740-6213-cfda-ecb340b5fa26" [ 945.078226] env[61986]: _type = "Task" [ 945.078226] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.086421] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52dcd6f6-1740-6213-cfda-ecb340b5fa26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.131165] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160080, 'name': ReconfigVM_Task, 'duration_secs': 1.147122} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.131471] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 945.131664] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 945.131922] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3975155f-83eb-401c-8139-3348edffd52d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.140774] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 945.140774] env[61986]: value = "task-1160083" [ 945.140774] env[61986]: _type = "Task" [ 945.140774] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.149997] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.155160] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160082, 'name': Destroy_Task, 'duration_secs': 0.48163} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.155160] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Destroyed the VM [ 945.156888] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 945.156888] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-01c05178-c206-4f3a-afd0-dc1b27612d53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.164893] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 945.164893] env[61986]: value = "task-1160084" [ 945.164893] env[61986]: _type = "Task" [ 945.164893] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.170350] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160084, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.407020] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573941} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.407398] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 945.407625] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.407897] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-deda16df-0397-4607-9415-0d97d0719fc4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.415898] env[61986]: DEBUG oslo_concurrency.lockutils [None req-024bb8f2-837a-41c4-b28d-dec9855ac1ea tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "076c8e91-af70-489f-89d5-f598166fafc6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.397s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.419704] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 945.419704] env[61986]: value = "task-1160085" [ 945.419704] env[61986]: _type = "Task" [ 945.419704] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.436953] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160085, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.467109] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.592796] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52dcd6f6-1740-6213-cfda-ecb340b5fa26, 'name': SearchDatastore_Task, 'duration_secs': 0.01723} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.594128] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5765b478-7fa7-4a71-98f8-39f68ec62fb1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.600929] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 945.600929] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fdf3b8-b1d9-e045-828c-0e23c999a901" [ 945.600929] env[61986]: _type = "Task" [ 945.600929] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.609268] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fdf3b8-b1d9-e045-828c-0e23c999a901, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.654669] env[61986]: DEBUG oslo_vmware.api [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160083, 'name': PowerOnVM_Task, 'duration_secs': 0.434039} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.654960] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 945.655230] env[61986]: DEBUG nova.compute.manager [None req-7abafa6d-6190-48a1-9ede-f8eb7b7d2aa8 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.656049] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3ac614-3a47-42e8-af37-f1252b3d3f18 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.670021] env[61986]: DEBUG nova.network.neutron [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [{"id": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "address": "fa:16:3e:5e:4f:21", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7c53e1f-f6", "ovs_interfaceid": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.679476] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160084, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.816301] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccff0a88-e772-4afd-81fb-8e8c150c86b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.825116] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affc59e5-6fc0-46d5-9125-d4c85e03c194 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.857950] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce59963-3d5d-4cdf-8fc7-2fee4d7b9828 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.866348] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3373a685-60d5-4e4f-b611-d8ace0afa360 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.882292] env[61986]: DEBUG nova.compute.provider_tree [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.934478] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160085, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07497} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.934764] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.936554] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6abf95c0-c306-4998-b677-02e7bb4b43d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.958365] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.961141] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94a5f06f-71e3-4baf-b441-36f91f409170 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.980753] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 945.980753] env[61986]: value = "task-1160086" [ 945.980753] env[61986]: _type = "Task" [ 945.980753] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.991818] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160086, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.111672] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fdf3b8-b1d9-e045-828c-0e23c999a901, 'name': SearchDatastore_Task, 'duration_secs': 0.011387} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.111956] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.112230] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] bc93bdd5-256e-4900-a9ae-609b0a7d402d/bc93bdd5-256e-4900-a9ae-609b0a7d402d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 946.112991] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1b3e574-5745-4be5-9791-88d5a39d0241 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.119248] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 946.119248] env[61986]: value = "task-1160087" [ 946.119248] env[61986]: _type = "Task" [ 946.119248] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.127625] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160087, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.172774] env[61986]: DEBUG oslo_concurrency.lockutils [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.177214] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160084, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.385946] env[61986]: DEBUG nova.scheduler.client.report [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.492505] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.631318] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160087, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.675550] env[61986]: DEBUG oslo_vmware.api [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160084, 'name': RemoveSnapshot_Task, 'duration_secs': 1.409819} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.675844] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 946.676113] env[61986]: INFO nova.compute.manager [None req-ae24d116-053e-4734-a586-683a89da0f64 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Took 19.58 seconds to snapshot the instance on the hypervisor. [ 946.713189] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c03e6f4-cf9e-499a-afd5-026e4c4438a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.739297] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cce889-4e98-4b10-9715-f3af7e13cc04 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.748928] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 83 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 946.892047] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.030s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.895417] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.333s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.895417] env[61986]: DEBUG nova.objects.instance [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lazy-loading 'resources' on Instance uuid 6e0697d4-d862-497c-9bdd-dd1ef2d4272b {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.921110] env[61986]: INFO nova.scheduler.client.report [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Deleted allocations for instance 52eda388-05f4-416c-ac9d-bd1c1e31ba9c [ 946.959516] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.959516] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.959516] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.959516] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.959516] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.961505] env[61986]: INFO nova.compute.manager [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Terminating instance [ 946.964043] env[61986]: DEBUG nova.compute.manager [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.964472] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 946.965436] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be53957-bbf9-4ff6-a474-71b9f8f56ab7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.975616] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 946.976062] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4c74624-7ead-47cd-8eb8-e4290c0d49fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.990176] env[61986]: DEBUG oslo_vmware.api [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 946.990176] env[61986]: value = "task-1160088" [ 946.990176] env[61986]: _type = "Task" [ 946.990176] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.991854] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160086, 'name': ReconfigVM_Task, 'duration_secs': 0.621092} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.995380] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Reconfigured VM instance instance-00000026 to attach disk [datastore2] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.996403] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fcee02de-ed17-4fae-bd27-a51a308613ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.004123] env[61986]: DEBUG oslo_vmware.api [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160088, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.005631] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 947.005631] env[61986]: value = "task-1160089" [ 947.005631] env[61986]: _type = "Task" [ 947.005631] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.017642] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160089, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.132806] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160087, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.665975} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.132806] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] bc93bdd5-256e-4900-a9ae-609b0a7d402d/bc93bdd5-256e-4900-a9ae-609b0a7d402d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 947.132806] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.133837] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ba033ec-1333-4188-94d2-5eb17eede15b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.142020] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 947.142020] env[61986]: value = "task-1160090" [ 947.142020] env[61986]: _type = "Task" [ 947.142020] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.149035] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160090, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.260284] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.260649] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed01ada5-729b-4ad3-97d4-ed0aa55c8e03 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.268871] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 947.268871] env[61986]: value = "task-1160091" [ 947.268871] env[61986]: _type = "Task" [ 947.268871] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.277748] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160091, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.434772] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d05db877-cdc9-43e5-b1b6-42860626989d tempest-ListServersNegativeTestJSON-372743732 tempest-ListServersNegativeTestJSON-372743732-project-member] Lock "52eda388-05f4-416c-ac9d-bd1c1e31ba9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.288s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.501389] env[61986]: DEBUG oslo_vmware.api [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160088, 'name': PowerOffVM_Task, 'duration_secs': 0.264435} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.504436] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.504784] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 947.505493] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c2439ae-464b-43e0-b10d-65fa1424c65f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.517989] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160089, 'name': Rename_Task, 'duration_secs': 0.142079} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.519159] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.520024] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c56be871-5ede-461b-9e88-592788a0e8ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.529975] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 947.529975] env[61986]: value = "task-1160093" [ 947.529975] env[61986]: _type = "Task" [ 947.529975] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.537453] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.650978] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160090, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063033} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.652174] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.653078] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31294bc1-6391-46a1-b597-5e320d68956e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.680053] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] bc93bdd5-256e-4900-a9ae-609b0a7d402d/bc93bdd5-256e-4900-a9ae-609b0a7d402d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.683008] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ec439aa-bd74-4aa6-b4d0-c70cd438f6c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.703662] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 947.703662] env[61986]: value = "task-1160094" [ 947.703662] env[61986]: _type = "Task" [ 947.703662] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.716660] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160094, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.779444] env[61986]: DEBUG oslo_vmware.api [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160091, 'name': PowerOnVM_Task, 'duration_secs': 0.412655} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.779728] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 947.779914] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-31c5adba-5cd2-47f3-995a-c5f8bfc6c3f5 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance 'f42e92ef-8a90-473c-8662-57cbb9e3f4ee' progress to 100 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 947.874112] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64c7a0d-df31-4f7d-8ca1-2f7955b97a56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.881944] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34d9637-f0ff-4f38-89a7-ed757046ccd6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.912325] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f371e5-6479-480d-8b88-1451699a257d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.920129] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b487ecd-5106-4e60-b4f9-aee9a599a0b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.935020] env[61986]: DEBUG nova.compute.provider_tree [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.039516] env[61986]: DEBUG oslo_vmware.api [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160093, 'name': PowerOnVM_Task, 'duration_secs': 0.496134} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.040109] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 948.040109] env[61986]: DEBUG nova.compute.manager [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.040816] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e099c2c-065c-4a21-8ad7-2b3ce983f1a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.213410] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.438601] env[61986]: DEBUG nova.scheduler.client.report [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.556462] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.715459] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160094, 'name': ReconfigVM_Task, 'duration_secs': 0.946207} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.716077] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Reconfigured VM instance instance-00000051 to attach disk [datastore2] bc93bdd5-256e-4900-a9ae-609b0a7d402d/bc93bdd5-256e-4900-a9ae-609b0a7d402d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.716501] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8aa60edf-871c-4462-a365-008acdaaf065 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.722275] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 948.722275] env[61986]: value = "task-1160095" [ 948.722275] env[61986]: _type = "Task" [ 948.722275] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.730828] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160095, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.946756] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.052s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.950416] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.133s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.954017] env[61986]: INFO nova.compute.claims [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.991102] env[61986]: INFO nova.scheduler.client.report [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Deleted allocations for instance 6e0697d4-d862-497c-9bdd-dd1ef2d4272b [ 949.233561] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160095, 'name': Rename_Task, 'duration_secs': 0.283835} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.233851] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 949.234119] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f27a0ee6-9dd2-448d-aaf2-3e7430307eac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.241779] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 949.241779] env[61986]: value = "task-1160096" [ 949.241779] env[61986]: _type = "Task" [ 949.241779] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.250628] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160096, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.499289] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6aed6f74-0b9d-41ee-8568-86a7f20f3a69 tempest-ServersAdminNegativeTestJSON-254054608 tempest-ServersAdminNegativeTestJSON-254054608-project-member] Lock "6e0697d4-d862-497c-9bdd-dd1ef2d4272b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.961s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.604348] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.604598] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.604791] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Deleting the datastore file [datastore1] 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.605176] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-add73e40-64fe-42c4-95b8-87913dfed068 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.612249] env[61986]: DEBUG oslo_vmware.api [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 949.612249] env[61986]: value = "task-1160097" [ 949.612249] env[61986]: _type = "Task" [ 949.612249] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.620977] env[61986]: DEBUG oslo_vmware.api [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160097, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.753534] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160096, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.774675] env[61986]: INFO nova.compute.manager [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Rebuilding instance [ 949.847976] env[61986]: DEBUG nova.compute.manager [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.848940] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab0f001-8503-4e55-9366-0a13263ece6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.962764] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.962764] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.963199] env[61986]: DEBUG nova.compute.manager [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Going to confirm migration 2 {{(pid=61986) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 949.989781] env[61986]: DEBUG nova.compute.manager [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.990753] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40abe09e-acd7-4815-905f-0398e572e1c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.122701] env[61986]: DEBUG oslo_vmware.api [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160097, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15329} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.125098] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.125299] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.125482] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.125748] env[61986]: INFO nova.compute.manager [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Took 3.16 seconds to destroy the instance on the hypervisor. [ 950.126013] env[61986]: DEBUG oslo.service.loopingcall [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.126393] env[61986]: DEBUG nova.compute.manager [-] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.126494] env[61986]: DEBUG nova.network.neutron [-] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.260493] env[61986]: DEBUG oslo_vmware.api [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160096, 'name': PowerOnVM_Task, 'duration_secs': 0.574556} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.260777] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.260983] env[61986]: INFO nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Took 9.63 seconds to spawn the instance on the hypervisor. [ 950.261184] env[61986]: DEBUG nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.261980] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b04e21-7ecc-40de-bc69-ef570f12a73b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.312324] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1c48d3-0457-46ab-8563-786f3a196cb6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.326443] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854ea7f8-1454-4ce3-b1b8-8af7d159d470 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.365316] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.365869] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5c4e8d1-f820-41b3-8f9e-b8fefe1a4880 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.368129] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee65d383-b4dd-4dfc-a2dd-4c4dcf37e79d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.377279] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ba3151-78f6-4c8e-96dd-d3d8488c8e5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.381658] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 950.381658] env[61986]: value = "task-1160098" [ 950.381658] env[61986]: _type = "Task" [ 950.381658] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.394069] env[61986]: DEBUG nova.compute.provider_tree [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.400852] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.509833] env[61986]: INFO nova.compute.manager [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] instance snapshotting [ 950.516876] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fc057d-bc83-4c51-8889-c7216b3fb5ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.538169] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b06fd4d-7456-43fc-bbe3-cc79697f0ed0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.584820] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.585075] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.585513] env[61986]: DEBUG nova.network.neutron [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.585513] env[61986]: DEBUG nova.objects.instance [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lazy-loading 'info_cache' on Instance uuid f42e92ef-8a90-473c-8662-57cbb9e3f4ee {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.786165] env[61986]: INFO nova.compute.manager [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Took 38.02 seconds to build instance. [ 950.803825] env[61986]: DEBUG nova.compute.manager [req-94d57f8e-15ec-4c4f-994d-9a8cb2c07e4d req-961cc2d4-cfea-40b0-82b9-41356fb9661c service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Received event network-vif-deleted-c4411312-86a5-48dd-a108-5c5bc29d5f3e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 950.803825] env[61986]: INFO nova.compute.manager [req-94d57f8e-15ec-4c4f-994d-9a8cb2c07e4d req-961cc2d4-cfea-40b0-82b9-41356fb9661c service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Neutron deleted interface c4411312-86a5-48dd-a108-5c5bc29d5f3e; detaching it from the instance and deleting it from the info cache [ 950.804030] env[61986]: DEBUG nova.network.neutron [req-94d57f8e-15ec-4c4f-994d-9a8cb2c07e4d req-961cc2d4-cfea-40b0-82b9-41356fb9661c service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.893460] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160098, 'name': PowerOffVM_Task, 'duration_secs': 0.195953} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.893938] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.894352] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.896600] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b8ab90-7516-456f-989f-b673ac76fc0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.901176] env[61986]: DEBUG nova.scheduler.client.report [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.919162] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.919162] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e853d73-ed9b-4bb8-bdd0-46255c65b89d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.978095] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.978472] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.978576] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleting the datastore file [datastore2] aa12919e-3891-4da9-a280-4155137864e1 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.978790] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-647e12d4-706d-4b8c-a328-82f9cadf2e1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.989330] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 950.989330] env[61986]: value = "task-1160100" [ 950.989330] env[61986]: _type = "Task" [ 950.989330] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.994683] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.049295] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 951.049663] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9f018230-5fa9-4cd5-b6d5-b40c13187a22 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.057311] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 951.057311] env[61986]: value = "task-1160101" [ 951.057311] env[61986]: _type = "Task" [ 951.057311] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.066528] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160101, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.163841] env[61986]: DEBUG nova.network.neutron [-] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.288812] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f59f15f5-c10d-45a8-91fe-562a3e6a8435 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.533s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.307418] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46f19eaa-3d21-4701-8ff6-b17dba69f9f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.317367] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7a9506-c071-40e6-bd56-d2e80afa3ffd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.356877] env[61986]: DEBUG nova.compute.manager [req-94d57f8e-15ec-4c4f-994d-9a8cb2c07e4d req-961cc2d4-cfea-40b0-82b9-41356fb9661c service nova] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Detach interface failed, port_id=c4411312-86a5-48dd-a108-5c5bc29d5f3e, reason: Instance 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 951.410327] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.412017] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.415519] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 31.102s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.496283] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133564} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.496619] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.496852] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 951.497086] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 951.566559] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160101, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.668331] env[61986]: INFO nova.compute.manager [-] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Took 1.54 seconds to deallocate network for instance. [ 951.904364] env[61986]: DEBUG nova.network.neutron [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [{"id": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "address": "fa:16:3e:5e:4f:21", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7c53e1f-f6", "ovs_interfaceid": "e7c53e1f-f6f4-4df9-b4b0-e2818f87b045", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.920843] env[61986]: DEBUG nova.compute.utils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.939753] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.940141] env[61986]: DEBUG nova.network.neutron [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 952.028793] env[61986]: DEBUG nova.policy [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4b4b1412f644b6097f4de846388dec1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691d4935610e45e6a0bae5ca5f1197f3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 952.070735] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160101, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.178337] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.274684] env[61986]: DEBUG nova.compute.manager [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 952.274940] env[61986]: DEBUG nova.compute.manager [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing instance network info cache due to event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 952.278863] env[61986]: DEBUG oslo_concurrency.lockutils [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.279087] env[61986]: DEBUG oslo_concurrency.lockutils [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.279273] env[61986]: DEBUG nova.network.neutron [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 952.406861] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-f42e92ef-8a90-473c-8662-57cbb9e3f4ee" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.407195] env[61986]: DEBUG nova.objects.instance [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lazy-loading 'migration_context' on Instance uuid f42e92ef-8a90-473c-8662-57cbb9e3f4ee {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.440127] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.443719] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Applying migration context for instance f42e92ef-8a90-473c-8662-57cbb9e3f4ee as it has an incoming, in-progress migration b772f020-d14b-4ce6-a2c7-eb4d645079b2. Migration status is confirming {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 952.445760] env[61986]: INFO nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating resource usage from migration f0c082cb-b5ab-49a7-a3ba-9675c6afa79d [ 952.445918] env[61986]: INFO nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating resource usage from migration b772f020-d14b-4ce6-a2c7-eb4d645079b2 [ 952.470069] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa12919e-3891-4da9-a280-4155137864e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.470237] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.470385] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance eaa61f81-2f9b-4d1c-bab8-4363fb71e936 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 952.470520] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 86c9e45c-b3f1-4004-9acc-190b11a4a926 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 952.470648] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9e57f8f9-6e9e-45fb-91d1-132490e930ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.470757] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ee47cab9-5dfd-48ce-ba70-cb800d735b19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.470979] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 021a27ec-cee0-454d-8daf-e6a82bd9330c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.471096] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a0fa2cbe-d8c3-462a-9744-aaa36a8d314e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 952.471188] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance b9706e51-0f74-4dbc-aab1-ea640ac78c62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.471368] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 59cfee64-ef79-4b8f-a703-c8812551d12f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.471533] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6488f68e-7b84-4462-aef5-25d02db504f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.471689] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 49bcac13-6dde-4a28-9131-f9a1e1d2e386 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 952.471815] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance dea128a9-aed0-40b2-ae17-c068ea8e3452 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 952.471930] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.472615] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 3d5f4513-bbc4-404e-9d3e-340bd369fc3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.472772] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 5170aa51-3307-42b1-b0dd-645dd4036e5b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.472913] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance f15bf19d-b86a-4b0a-ac1a-9df8e77e9382 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.473058] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance f9aa0511-b05c-408f-ac06-c49bf8dc648c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 952.473306] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Migration b772f020-d14b-4ce6-a2c7-eb4d645079b2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 952.474065] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance f42e92ef-8a90-473c-8662-57cbb9e3f4ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.474768] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance bc93bdd5-256e-4900-a9ae-609b0a7d402d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.475314] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 221e10a3-da31-410c-80f8-4bcc2c515710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.537540] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.537819] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.538059] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.538325] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.538504] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.538933] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.538933] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.539121] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.539243] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.539468] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.539695] env[61986]: DEBUG nova.virt.hardware [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.543131] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29b2d42-7366-45de-90cd-b6b48484ff1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.549420] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292a9812-c8dc-4cd7-9235-e6bae84e2521 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.565757] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:61:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '038ca95f-90ee-4bd7-afb9-b514378d344e', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.574682] env[61986]: DEBUG oslo.service.loopingcall [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.580105] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 952.580455] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c203fbd8-70da-499c-94e4-706ed9420370 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.601228] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160101, 'name': CreateSnapshot_Task, 'duration_secs': 1.276074} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.602696] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 952.602951] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.602951] env[61986]: value = "task-1160102" [ 952.602951] env[61986]: _type = "Task" [ 952.602951] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.603674] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7424ce25-5278-4855-95ec-09b77b9424a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.623764] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160102, 'name': CreateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.898556] env[61986]: DEBUG nova.network.neutron [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Successfully created port: a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.910171] env[61986]: DEBUG nova.objects.base [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 952.911292] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0048ffec-1099-4f1c-bf86-f153aa306902 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.932141] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-969814af-9f5e-420b-bfe9-e120f094d545 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.937964] env[61986]: DEBUG oslo_vmware.api [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 952.937964] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d08e66-339a-8ca0-60a1-02f016fc492f" [ 952.937964] env[61986]: _type = "Task" [ 952.937964] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.950095] env[61986]: DEBUG oslo_vmware.api [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d08e66-339a-8ca0-60a1-02f016fc492f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.986895] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 3b925174-cf0b-4c6c-b6b6-1c99a50026be has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 952.988264] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Migration f0c082cb-b5ab-49a7-a3ba-9675c6afa79d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 952.988434] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.988704] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 952.988844] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3840MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=17 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 953.135393] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 953.135759] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160102, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.136091] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-80890c7d-0261-4861-a4de-83fffa9d7430 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.149035] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 953.149035] env[61986]: value = "task-1160103" [ 953.149035] env[61986]: _type = "Task" [ 953.149035] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.162969] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160103, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.184991] env[61986]: DEBUG nova.network.neutron [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated VIF entry in instance network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 953.185378] env[61986]: DEBUG nova.network.neutron [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.453274] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.455945] env[61986]: DEBUG oslo_vmware.api [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d08e66-339a-8ca0-60a1-02f016fc492f, 'name': SearchDatastore_Task, 'duration_secs': 0.009745} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.458873] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.474673] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e786ca-5b06-4bce-9009-dc654a2df627 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.490769] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.491131] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.491334] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.491530] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.491835] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.492096] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.492342] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.492576] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.492888] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.493092] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.493417] env[61986]: DEBUG nova.virt.hardware [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.495308] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff0d360-1b17-47aa-a8e6-8c107cf9c37a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.502121] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163b4336-fb09-4d5d-ae51-0a5bcd47f982 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.533612] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec823f31-b41a-4f89-81e2-64019b241962 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.538287] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eaec89-1f83-4aee-b7a9-d055ccb2ab64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.553951] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574bff66-2e8b-4d7d-8968-1a662d2e8943 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.569118] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.622661] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160102, 'name': CreateVM_Task, 'duration_secs': 0.579421} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.622825] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 953.623845] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.623845] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.624079] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.624284] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39658010-6aea-494c-b0d3-126ace0a1b48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.632338] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 953.632338] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e17e54-7b0e-0f36-85b1-3273a3507d6a" [ 953.632338] env[61986]: _type = "Task" [ 953.632338] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.644785] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e17e54-7b0e-0f36-85b1-3273a3507d6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.659124] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160103, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.689048] env[61986]: DEBUG oslo_concurrency.lockutils [req-9484eb49-efb5-4896-9d91-eabcb6fe0533 req-ad774622-b64f-4c34-9fd2-9a8a48c80fe0 service nova] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.733819] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.734155] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bda7750f-eec9-40d9-ace5-18d48234126e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.072684] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.146779] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e17e54-7b0e-0f36-85b1-3273a3507d6a, 'name': SearchDatastore_Task, 'duration_secs': 0.01354} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.147161] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.147791] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.147791] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.147791] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.148029] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.150386] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ccfbcca-9dd7-48a4-8522-8a3a738d05bb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.160792] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160103, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.161918] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.162107] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 954.162866] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4d1b94c-f121-40f4-b85e-89d16c5fe8be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.167810] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 954.167810] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e33440-13ef-f67d-4d8b-9fe481cb0b5a" [ 954.167810] env[61986]: _type = "Task" [ 954.167810] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.177969] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e33440-13ef-f67d-4d8b-9fe481cb0b5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.237986] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.583505] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 954.583505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.165s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.583505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.973s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.583505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.583505] env[61986]: INFO nova.compute.manager [None req-c0921785-0d14-47e4-a174-00cd3452522c tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Successfully reverted task state from rebuilding on failure for instance. [ 954.589985] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.309s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.590450] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.592970] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.947s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.593329] env[61986]: DEBUG nova.objects.instance [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 954.665768] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160103, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.679750] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e33440-13ef-f67d-4d8b-9fe481cb0b5a, 'name': SearchDatastore_Task, 'duration_secs': 0.009402} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.680720] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5336a9be-5994-47b1-a514-809278fcf7a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.686566] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 954.686566] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5273b853-1850-c893-431d-5139af488b72" [ 954.686566] env[61986]: _type = "Task" [ 954.686566] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.695694] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5273b853-1850-c893-431d-5139af488b72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.765132] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.989672] env[61986]: DEBUG nova.network.neutron [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Successfully updated port: a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.050870] env[61986]: DEBUG nova.compute.manager [req-ee64365d-b35e-47a5-8d42-f276e4ebef8e req-82397355-b8e5-4661-b606-814e82871a66 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-vif-plugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.050870] env[61986]: DEBUG oslo_concurrency.lockutils [req-ee64365d-b35e-47a5-8d42-f276e4ebef8e req-82397355-b8e5-4661-b606-814e82871a66 service nova] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.050870] env[61986]: DEBUG oslo_concurrency.lockutils [req-ee64365d-b35e-47a5-8d42-f276e4ebef8e req-82397355-b8e5-4661-b606-814e82871a66 service nova] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.050870] env[61986]: DEBUG oslo_concurrency.lockutils [req-ee64365d-b35e-47a5-8d42-f276e4ebef8e req-82397355-b8e5-4661-b606-814e82871a66 service nova] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.051616] env[61986]: DEBUG nova.compute.manager [req-ee64365d-b35e-47a5-8d42-f276e4ebef8e req-82397355-b8e5-4661-b606-814e82871a66 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] No waiting events found dispatching network-vif-plugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.052036] env[61986]: WARNING nova.compute.manager [req-ee64365d-b35e-47a5-8d42-f276e4ebef8e req-82397355-b8e5-4661-b606-814e82871a66 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received unexpected event network-vif-plugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 for instance with vm_state building and task_state spawning. [ 955.124121] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1e0048c-e554-4ac1-80d3-16260f719ebd tempest-ServerActionsV293TestJSON-1028145175 tempest-ServerActionsV293TestJSON-1028145175-project-member] Lock "98337bb6-9502-4d4c-af00-028659b246bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.879s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.162750] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160103, 'name': CloneVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.199440] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5273b853-1850-c893-431d-5139af488b72, 'name': SearchDatastore_Task, 'duration_secs': 0.009726} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.200646] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.202991] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 955.202991] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1659c955-0854-44c3-9399-5f8026611177 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.209640] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 955.209640] env[61986]: value = "task-1160104" [ 955.209640] env[61986]: _type = "Task" [ 955.209640] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.219683] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160104, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.494063] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.495719] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.495719] env[61986]: DEBUG nova.network.neutron [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.590750] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.590750] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.611903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc3f564-0fae-47b5-b2eb-7ddd1a8b916b tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.611903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.921s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.611903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.615320] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.780s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.615539] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.617701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.007s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.617896] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.619793] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.021s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.621761] env[61986]: INFO nova.compute.claims [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.664924] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160103, 'name': CloneVM_Task, 'duration_secs': 2.072801} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.664924] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Created linked-clone VM from snapshot [ 955.666185] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bacf502-2a92-44f2-8776-6437000a2401 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.676452] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Uploading image db3e6c6f-cc3d-4ae3-90f2-666514c372e9 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 955.686698] env[61986]: INFO nova.scheduler.client.report [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Deleted allocations for instance dea128a9-aed0-40b2-ae17-c068ea8e3452 [ 955.691324] env[61986]: INFO nova.scheduler.client.report [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Deleted allocations for instance 86c9e45c-b3f1-4004-9acc-190b11a4a926 [ 955.709397] env[61986]: INFO nova.scheduler.client.report [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleted allocations for instance 49bcac13-6dde-4a28-9131-f9a1e1d2e386 [ 955.714072] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 955.714072] env[61986]: value = "vm-252412" [ 955.714072] env[61986]: _type = "VirtualMachine" [ 955.714072] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 955.714072] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0d05ebf8-9459-47a5-a95d-0bd70e44026d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.729595] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160104, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512624} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.731015] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 955.731469] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.731825] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lease: (returnval){ [ 955.731825] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d5afff-df56-bc7e-8f47-0f78282cc4d8" [ 955.731825] env[61986]: _type = "HttpNfcLease" [ 955.731825] env[61986]: } obtained for exporting VM: (result){ [ 955.731825] env[61986]: value = "vm-252412" [ 955.731825] env[61986]: _type = "VirtualMachine" [ 955.731825] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 955.732113] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the lease: (returnval){ [ 955.732113] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d5afff-df56-bc7e-8f47-0f78282cc4d8" [ 955.732113] env[61986]: _type = "HttpNfcLease" [ 955.732113] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 955.732264] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d288dda-605a-4870-a815-ab22f70dc11e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.745880] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 955.745880] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d5afff-df56-bc7e-8f47-0f78282cc4d8" [ 955.745880] env[61986]: _type = "HttpNfcLease" [ 955.745880] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 955.747618] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 955.747618] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d5afff-df56-bc7e-8f47-0f78282cc4d8" [ 955.747618] env[61986]: _type = "HttpNfcLease" [ 955.747618] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 955.747618] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 955.747618] env[61986]: value = "task-1160106" [ 955.747618] env[61986]: _type = "Task" [ 955.747618] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.748325] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e615053-9086-4c61-8f25-6cc86dfdc58c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.760750] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293012a-5588-051f-3090-b6e4056aa571/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 955.761052] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293012a-5588-051f-3090-b6e4056aa571/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 955.767483] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160106, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.878739] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fc6f4cea-26bd-4849-b9f7-3871f788875c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.034956] env[61986]: DEBUG nova.network.neutron [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.104295] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.104490] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 956.206420] env[61986]: DEBUG oslo_concurrency.lockutils [None req-82f175da-8eae-46e4-8d99-339185157715 tempest-FloatingIPsAssociationTestJSON-174490616 tempest-FloatingIPsAssociationTestJSON-174490616-project-member] Lock "86c9e45c-b3f1-4004-9acc-190b11a4a926" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.438s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.208036] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ad91c3ce-f4f6-4b76-8244-786ca05efc7d tempest-ServerShowV254Test-422165862 tempest-ServerShowV254Test-422165862-project-member] Lock "dea128a9-aed0-40b2-ae17-c068ea8e3452" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.340s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.225419] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cbb75000-b8a5-4e9a-a815-39b0e8c3d0f5 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "49bcac13-6dde-4a28-9131-f9a1e1d2e386" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.872s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.261610] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160106, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096726} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.262311] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.263434] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc17534c-069b-4a06-b690-587813870dc7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.301094] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.302326] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d5c37ed-3e9c-4c70-be9b-8f8cb0e6f414 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.325255] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 956.325255] env[61986]: value = "task-1160107" [ 956.325255] env[61986]: _type = "Task" [ 956.325255] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.336952] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160107, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.367977] env[61986]: DEBUG nova.network.neutron [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.414912] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.415216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.655770] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.655948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.656121] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 956.839707] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160107, 'name': ReconfigVM_Task, 'duration_secs': 0.299365} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.843951] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Reconfigured VM instance instance-00000026 to attach disk [datastore1] aa12919e-3891-4da9-a280-4155137864e1/aa12919e-3891-4da9-a280-4155137864e1.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.845173] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19179362-880f-4242-934e-71c5a753ec62 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.853137] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 956.853137] env[61986]: value = "task-1160108" [ 956.853137] env[61986]: _type = "Task" [ 956.853137] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.868853] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160108, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.875127] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.875509] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance network_info: |[{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 956.876357] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:9b:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1768af3d-3317-4ef5-b484-0c2707d63de7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.884548] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating folder: Project (691d4935610e45e6a0bae5ca5f1197f3). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 956.884762] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f90aafd-384e-4143-9e93-089f7d96b560 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.898040] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created folder: Project (691d4935610e45e6a0bae5ca5f1197f3) in parent group-v252271. [ 956.898697] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating folder: Instances. Parent ref: group-v252413. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 956.898958] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34568d5b-fabf-4981-b8c0-06f3282d4f7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.908254] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created folder: Instances in parent group-v252413. [ 956.908749] env[61986]: DEBUG oslo.service.loopingcall [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.909098] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 956.909342] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e31c613-7d3e-4a19-9589-ae3d3e062fde {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.928840] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.941087] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.941087] env[61986]: value = "task-1160111" [ 956.941087] env[61986]: _type = "Task" [ 956.941087] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.954421] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160111, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.053803] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2521879f-6ec5-40ec-aa63-7bb9a565d2e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.061607] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0239bd95-9060-48c5-b290-2d0453ced71a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.098730] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11e343e-a744-4e09-9c55-5e13c0835d14 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.106680] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54478aa-c4a6-41bf-be7e-c2f105450119 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.120855] env[61986]: DEBUG nova.compute.provider_tree [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.312333] env[61986]: DEBUG nova.compute.manager [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.312999] env[61986]: DEBUG nova.compute.manager [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing instance network info cache due to event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.313248] env[61986]: DEBUG oslo_concurrency.lockutils [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.313415] env[61986]: DEBUG oslo_concurrency.lockutils [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.313586] env[61986]: DEBUG nova.network.neutron [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.367692] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "73b2c463-0460-4a68-b788-ab0c7818d8f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.367692] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.373716] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160108, 'name': Rename_Task, 'duration_secs': 0.165933} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.374375] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.374691] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87cac387-cd13-4901-8aaf-d3c98ffcf75b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.383144] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 957.383144] env[61986]: value = "task-1160112" [ 957.383144] env[61986]: _type = "Task" [ 957.383144] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.397238] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160112, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.452546] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160111, 'name': CreateVM_Task, 'duration_secs': 0.470309} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.452546] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 957.452861] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.453155] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.454029] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.454080] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3ee1f54-98e6-4fc5-b542-21dd68f57727 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.460078] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 957.460078] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5291e144-1f6c-3f43-a2a9-65eb3cc2ffcd" [ 957.460078] env[61986]: _type = "Task" [ 957.460078] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.466958] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.476283] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5291e144-1f6c-3f43-a2a9-65eb3cc2ffcd, 'name': SearchDatastore_Task, 'duration_secs': 0.013179} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.476711] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.477085] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.477465] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.477690] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.477943] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.478281] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fea80240-3773-40ad-9293-7874867e00e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.486810] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.487223] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 957.488284] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-446394e0-a20b-467c-9a60-df92dec3f5fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.495268] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 957.495268] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523ae02d-68dd-5022-9a51-4c3432587ee8" [ 957.495268] env[61986]: _type = "Task" [ 957.495268] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.504654] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523ae02d-68dd-5022-9a51-4c3432587ee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.627199] env[61986]: DEBUG nova.scheduler.client.report [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.874737] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 957.898695] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160112, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.015835] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523ae02d-68dd-5022-9a51-4c3432587ee8, 'name': SearchDatastore_Task, 'duration_secs': 0.011784} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.016825] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90685741-bd4f-4da5-9bf0-55b079a2dd07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.029959] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 958.029959] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d61463-d590-f463-a5bd-0a10db24fa0a" [ 958.029959] env[61986]: _type = "Task" [ 958.029959] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.039346] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d61463-d590-f463-a5bd-0a10db24fa0a, 'name': SearchDatastore_Task, 'duration_secs': 0.009334} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.039658] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.039950] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 958.040236] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a563e87-2457-4a83-92e4-fc5412fc857f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.046766] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 958.046766] env[61986]: value = "task-1160113" [ 958.046766] env[61986]: _type = "Task" [ 958.046766] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.058027] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160113, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.080798] env[61986]: DEBUG nova.network.neutron [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updated VIF entry in instance network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 958.081297] env[61986]: DEBUG nova.network.neutron [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.135699] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.136374] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 958.139282] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.137s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.139610] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.141842] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.219s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.142211] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.146205] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.701s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.207598] env[61986]: INFO nova.scheduler.client.report [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Deleted allocations for instance a0fa2cbe-d8c3-462a-9744-aaa36a8d314e [ 958.211145] env[61986]: INFO nova.scheduler.client.report [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted allocations for instance eaa61f81-2f9b-4d1c-bab8-4363fb71e936 [ 958.403735] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160112, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.409072] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.535351] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updating instance_info_cache with network_info: [{"id": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "address": "fa:16:3e:6a:56:8b", "network": {"id": "e26a0bd9-52c8-4e74-8c12-3ae136fe2613", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1819414114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f746f57f466f480c958d8b15bd2ce8fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a2fed5-6f", "ovs_interfaceid": "d9a2fed5-6f97-43b2-9e7a-5c859c341ebe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.560808] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160113, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467411} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.561128] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 958.561339] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.561901] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a15c057-6096-4b93-9bae-10dfdce30662 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.571413] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 958.571413] env[61986]: value = "task-1160114" [ 958.571413] env[61986]: _type = "Task" [ 958.571413] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.583099] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160114, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.593358] env[61986]: DEBUG oslo_concurrency.lockutils [req-a718c4d9-63d2-49bd-ae5a-1d186f9e3c55 req-9f8aea0c-e34c-4377-8ef0-5c46d4619779 service nova] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.655064] env[61986]: INFO nova.compute.claims [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.663043] env[61986]: DEBUG nova.compute.utils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.664624] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.664803] env[61986]: DEBUG nova.network.neutron [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.724855] env[61986]: DEBUG oslo_concurrency.lockutils [None req-811fa42e-7826-48f6-a005-dfc799ae6188 tempest-AttachVolumeTestJSON-441536152 tempest-AttachVolumeTestJSON-441536152-project-member] Lock "a0fa2cbe-d8c3-462a-9744-aaa36a8d314e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.162s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.726241] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4db930f8-1de1-48d5-ab38-cbcbb3d35d63 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "eaa61f81-2f9b-4d1c-bab8-4363fb71e936" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.052s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.781392] env[61986]: DEBUG nova.policy [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc6881bdd48c4f0ea6bcb7d2d3746d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e9d0c35a4de4f5f9829a7f3c88fde92', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 958.896304] env[61986]: DEBUG oslo_vmware.api [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160112, 'name': PowerOnVM_Task, 'duration_secs': 1.130854} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.896656] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.897327] env[61986]: DEBUG nova.compute.manager [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.901222] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f6376f-3cf5-4b26-934d-e844e3783eef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.038215] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.040671] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 959.040671] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.040671] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.040671] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.040671] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.084257] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160114, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079455} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.084721] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.085771] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc215a21-f390-45e5-ab5e-32d41ace1359 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.113324] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.113324] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae86d529-e5a0-4afc-9283-5bf776a7d798 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.139412] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 959.139412] env[61986]: value = "task-1160115" [ 959.139412] env[61986]: _type = "Task" [ 959.139412] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.150312] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160115, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.169711] env[61986]: INFO nova.compute.resource_tracker [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating resource usage from migration f0c082cb-b5ab-49a7-a3ba-9675c6afa79d [ 959.174710] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 959.425351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.600023] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289f5307-9036-4e42-b053-30a68c40926b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.607101] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962ac3d3-3826-455d-a89e-53c8be84eb34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.612426] env[61986]: DEBUG nova.network.neutron [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Successfully created port: a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.654447] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1db095-2e42-49d2-b41e-b24434c39c65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.663117] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160115, 'name': ReconfigVM_Task, 'duration_secs': 0.50859} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.666651] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.666651] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01955c8a-1484-49f9-85ed-00567630be41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.670116] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8590b583-0b0e-4601-aed4-8530c3034567 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.688964] env[61986]: DEBUG nova.compute.provider_tree [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.692126] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 959.692126] env[61986]: value = "task-1160117" [ 959.692126] env[61986]: _type = "Task" [ 959.692126] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.704062] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160117, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.192836] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 960.196200] env[61986]: DEBUG nova.scheduler.client.report [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.217779] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160117, 'name': Rename_Task, 'duration_secs': 0.172618} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.218334] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 960.220873] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5433be55-8932-4958-9785-605ee43426b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.229217] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 960.229217] env[61986]: value = "task-1160118" [ 960.229217] env[61986]: _type = "Task" [ 960.229217] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.234812] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.235081] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.235242] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.235437] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.235587] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.235735] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.235943] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.236118] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.236283] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.236447] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.236622] env[61986]: DEBUG nova.virt.hardware [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.237731] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1bd032-1ce5-49a4-b3a0-f6e8065ad537 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.248511] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008ddcb2-12b6-4376-a4c0-78c913075a1d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.252712] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160118, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.707628] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.561s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.707938] env[61986]: INFO nova.compute.manager [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Migrating [ 960.714738] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.251s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.714944] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.717055] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.161s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.717250] env[61986]: DEBUG nova.objects.instance [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 960.739850] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.778454] env[61986]: INFO nova.scheduler.client.report [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Deleted allocations for instance f9aa0511-b05c-408f-ac06-c49bf8dc648c [ 961.233186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.233494] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.233849] env[61986]: DEBUG nova.network.neutron [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.250364] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160118, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.289966] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9af1c3a8-859f-44ba-8250-696b8e515bbd tempest-ImagesNegativeTestJSON-791325733 tempest-ImagesNegativeTestJSON-791325733-project-member] Lock "f9aa0511-b05c-408f-ac06-c49bf8dc648c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.869s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.653572] env[61986]: DEBUG nova.network.neutron [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Successfully updated port: a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.737515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-818df435-3b62-4801-ae0c-d48dd6e2af44 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.743194] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.564s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.743194] env[61986]: DEBUG nova.objects.instance [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lazy-loading 'resources' on Instance uuid 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.752851] env[61986]: DEBUG oslo_vmware.api [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160118, 'name': PowerOnVM_Task, 'duration_secs': 1.067331} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.752851] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 961.753051] env[61986]: INFO nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Took 8.30 seconds to spawn the instance on the hypervisor. [ 961.753268] env[61986]: DEBUG nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.754121] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6664f4fb-d478-41f1-9985-8e06952fe0cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.850557] env[61986]: DEBUG nova.compute.manager [req-60a350f9-e0ac-4497-84c3-a946dcf36995 req-0f27e62c-3e56-4490-87a5-8eb8908ac255 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Received event network-vif-plugged-a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.850840] env[61986]: DEBUG oslo_concurrency.lockutils [req-60a350f9-e0ac-4497-84c3-a946dcf36995 req-0f27e62c-3e56-4490-87a5-8eb8908ac255 service nova] Acquiring lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.851092] env[61986]: DEBUG oslo_concurrency.lockutils [req-60a350f9-e0ac-4497-84c3-a946dcf36995 req-0f27e62c-3e56-4490-87a5-8eb8908ac255 service nova] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.851281] env[61986]: DEBUG oslo_concurrency.lockutils [req-60a350f9-e0ac-4497-84c3-a946dcf36995 req-0f27e62c-3e56-4490-87a5-8eb8908ac255 service nova] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.851563] env[61986]: DEBUG nova.compute.manager [req-60a350f9-e0ac-4497-84c3-a946dcf36995 req-0f27e62c-3e56-4490-87a5-8eb8908ac255 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] No waiting events found dispatching network-vif-plugged-a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.851700] env[61986]: WARNING nova.compute.manager [req-60a350f9-e0ac-4497-84c3-a946dcf36995 req-0f27e62c-3e56-4490-87a5-8eb8908ac255 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Received unexpected event network-vif-plugged-a6a66909-ff92-4f28-a4c4-c3811691128d for instance with vm_state building and task_state spawning. [ 962.006569] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.007093] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.007546] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.008624] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.008624] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.010566] env[61986]: INFO nova.compute.manager [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Terminating instance [ 962.015768] env[61986]: DEBUG nova.compute.manager [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.016072] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 962.018025] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c07d4c2-706b-4745-90f6-03be0cd80b3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.028626] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 962.028949] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b91efbc9-505a-4619-bb78-542eaf057f4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.043225] env[61986]: DEBUG oslo_vmware.api [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 962.043225] env[61986]: value = "task-1160120" [ 962.043225] env[61986]: _type = "Task" [ 962.043225] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.052800] env[61986]: DEBUG oslo_vmware.api [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.156173] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.156348] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.156547] env[61986]: DEBUG nova.network.neutron [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 962.275659] env[61986]: INFO nova.compute.manager [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Took 42.49 seconds to build instance. [ 962.341301] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "45d80bab-0140-4473-9dc2-4a5481838aed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.341563] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "45d80bab-0140-4473-9dc2-4a5481838aed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.488164] env[61986]: DEBUG nova.network.neutron [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.558059] env[61986]: DEBUG oslo_vmware.api [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160120, 'name': PowerOffVM_Task, 'duration_secs': 0.211783} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.558420] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 962.558629] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 962.558943] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17e5f2d1-6896-459b-875f-0576a1312104 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.618987] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b891ae0a-8c10-4475-8921-3979c02662e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.630795] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d41f97a-d989-4b35-92db-53d00e031a2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.637268] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 962.637600] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 962.637863] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleting the datastore file [datastore1] 3d5f4513-bbc4-404e-9d3e-340bd369fc3c {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.638958] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eed43948-c64e-484c-9b12-ce92326686ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.687302] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1e9660-8f55-44d9-8399-3c827564660e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.693523] env[61986]: DEBUG oslo_vmware.api [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 962.693523] env[61986]: value = "task-1160122" [ 962.693523] env[61986]: _type = "Task" [ 962.693523] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.703516] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b770a87e-7acf-48a9-b03f-ac9890fd13a7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.714392] env[61986]: DEBUG oslo_vmware.api [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.730495] env[61986]: DEBUG nova.compute.provider_tree [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.746160] env[61986]: DEBUG nova.network.neutron [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 962.778482] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c3fc370f-892d-438c-9535-db9418df3abd tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.979s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.844255] env[61986]: DEBUG nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 962.992087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.012648] env[61986]: DEBUG nova.network.neutron [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Updating instance_info_cache with network_info: [{"id": "a6a66909-ff92-4f28-a4c4-c3811691128d", "address": "fa:16:3e:ff:fe:82", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a66909-ff", "ovs_interfaceid": "a6a66909-ff92-4f28-a4c4-c3811691128d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.204809] env[61986]: DEBUG oslo_vmware.api [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139902} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.205174] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.205382] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 963.205610] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 963.205804] env[61986]: INFO nova.compute.manager [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Took 1.19 seconds to destroy the instance on the hypervisor. [ 963.206103] env[61986]: DEBUG oslo.service.loopingcall [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.206336] env[61986]: DEBUG nova.compute.manager [-] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.206449] env[61986]: DEBUG nova.network.neutron [-] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 963.230350] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.230598] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.235028] env[61986]: DEBUG nova.scheduler.client.report [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.374369] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.515709] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.515888] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance network_info: |[{"id": "a6a66909-ff92-4f28-a4c4-c3811691128d", "address": "fa:16:3e:ff:fe:82", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a66909-ff", "ovs_interfaceid": "a6a66909-ff92-4f28-a4c4-c3811691128d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 963.516351] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:fe:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6a66909-ff92-4f28-a4c4-c3811691128d', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 963.524795] env[61986]: DEBUG oslo.service.loopingcall [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.525291] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 963.525700] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b2cf99e-f864-485a-9487-47fcd25a872f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.567151] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 963.567151] env[61986]: value = "task-1160123" [ 963.567151] env[61986]: _type = "Task" [ 963.567151] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.577156] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160123, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.735620] env[61986]: DEBUG nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.741605] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.999s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.746137] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.287s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.784855] env[61986]: INFO nova.scheduler.client.report [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Deleted allocations for instance 93c4f104-1812-4bb7-bfa7-cbf70a19ff51 [ 963.975391] env[61986]: DEBUG nova.compute.manager [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Received event network-changed-a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.975391] env[61986]: DEBUG nova.compute.manager [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Refreshing instance network info cache due to event network-changed-a6a66909-ff92-4f28-a4c4-c3811691128d. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.975391] env[61986]: DEBUG oslo_concurrency.lockutils [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] Acquiring lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.975391] env[61986]: DEBUG oslo_concurrency.lockutils [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] Acquired lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.975391] env[61986]: DEBUG nova.network.neutron [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Refreshing network info cache for port a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 964.081840] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160123, 'name': CreateVM_Task, 'duration_secs': 0.466227} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.082216] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 964.085019] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.085019] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.085019] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 964.085019] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c55333fe-e17e-4efe-815d-601ab9e6ac24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.091032] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 964.091032] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520c9c9a-a96f-f462-97b1-79b5fd88ffcf" [ 964.091032] env[61986]: _type = "Task" [ 964.091032] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.101233] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520c9c9a-a96f-f462-97b1-79b5fd88ffcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.277326] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.297547] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1bb28525-40c8-4cb6-8e39-01e09c297d6e tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "93c4f104-1812-4bb7-bfa7-cbf70a19ff51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.340s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.408225] env[61986]: DEBUG nova.network.neutron [-] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.513538] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f35af7a-e06d-4a34-887b-a8d9a82b4c20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.543614] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 964.607331] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520c9c9a-a96f-f462-97b1-79b5fd88ffcf, 'name': SearchDatastore_Task, 'duration_secs': 0.01016} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.612093] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.612720] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.612720] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.612857] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.612940] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.614109] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe344982-e757-41f7-bc4d-7fdd61bc7fbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.627602] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.627818] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 964.628679] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba2b390e-c185-4185-8312-0558ec06bdb7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.637234] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 964.637234] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5275eac7-12da-5fc3-fef1-8e4ad28da555" [ 964.637234] env[61986]: _type = "Task" [ 964.637234] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.650940] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5275eac7-12da-5fc3-fef1-8e4ad28da555, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.725859] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f1ad7f-6f0b-40df-b364-f709047d3f16 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.734728] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573aca56-7a53-46dd-827c-f54e134706ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.768360] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9445aa7-60b7-4d48-96b8-f0b882aeaac4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.776820] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee75f1db-36cc-476a-96c0-ae224fc2955f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.790957] env[61986]: DEBUG nova.compute.provider_tree [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.872878] env[61986]: DEBUG nova.network.neutron [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Updated VIF entry in instance network info cache for port a6a66909-ff92-4f28-a4c4-c3811691128d. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.873261] env[61986]: DEBUG nova.network.neutron [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Updating instance_info_cache with network_info: [{"id": "a6a66909-ff92-4f28-a4c4-c3811691128d", "address": "fa:16:3e:ff:fe:82", "network": {"id": "6265c99d-9c3d-47e2-aaed-a6d2fc92fd70", "bridge": "br-int", "label": "tempest-ImagesTestJSON-138670064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e9d0c35a4de4f5f9829a7f3c88fde92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6a66909-ff", "ovs_interfaceid": "a6a66909-ff92-4f28-a4c4-c3811691128d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.911789] env[61986]: INFO nova.compute.manager [-] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Took 1.71 seconds to deallocate network for instance. [ 965.051420] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 965.051769] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e79254a-8194-4f8a-87ea-2dd8416da2ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.061402] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 965.061402] env[61986]: value = "task-1160124" [ 965.061402] env[61986]: _type = "Task" [ 965.061402] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.069439] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.152574] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5275eac7-12da-5fc3-fef1-8e4ad28da555, 'name': SearchDatastore_Task, 'duration_secs': 0.014644} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.156331] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eab5453c-2bba-468e-8325-a54c4eeda08c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.172711] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 965.172711] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b048af-c108-9314-e3d9-d0e0dc8c60fe" [ 965.172711] env[61986]: _type = "Task" [ 965.172711] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.188466] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b048af-c108-9314-e3d9-d0e0dc8c60fe, 'name': SearchDatastore_Task, 'duration_secs': 0.011575} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.189574] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.189901] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3b925174-cf0b-4c6c-b6b6-1c99a50026be/3b925174-cf0b-4c6c-b6b6-1c99a50026be.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 965.190273] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9abdac19-0566-4153-83b9-99002216d738 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.201091] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 965.201091] env[61986]: value = "task-1160125" [ 965.201091] env[61986]: _type = "Task" [ 965.201091] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.214612] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.214834] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 965.215181] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.294067] env[61986]: DEBUG nova.scheduler.client.report [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.379148] env[61986]: DEBUG oslo_concurrency.lockutils [req-539e23d5-6a63-489c-82c5-837e737f37ed req-547b86bc-9a90-4cab-b6ab-cbe051569fb8 service nova] Releasing lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.422922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.575267] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 965.576134] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 965.667698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "021a27ec-cee0-454d-8daf-e6a82bd9330c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.668045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.668283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "021a27ec-cee0-454d-8daf-e6a82bd9330c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.669372] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.669372] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.671742] env[61986]: INFO nova.compute.manager [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Terminating instance [ 965.677890] env[61986]: DEBUG nova.compute.manager [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 965.677890] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 965.677890] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a283f1-9b85-4686-8edc-fcd5c463f432 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.688963] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 965.689294] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4c6e4c0-58ed-45f5-bb97-963a6498862a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.700379] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 965.700379] env[61986]: value = "task-1160126" [ 965.700379] env[61986]: _type = "Task" [ 965.700379] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.719963] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.723486] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160125, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.812140] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293012a-5588-051f-3090-b6e4056aa571/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 965.813353] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b46663-eed1-4b18-84cb-ef1972a2ecc1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.821247] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293012a-5588-051f-3090-b6e4056aa571/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 965.821349] env[61986]: ERROR oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293012a-5588-051f-3090-b6e4056aa571/disk-0.vmdk due to incomplete transfer. [ 965.822562] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-06e677b9-ebac-4379-a4d0-c76f13c7df13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.835550] env[61986]: DEBUG oslo_vmware.rw_handles [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293012a-5588-051f-3090-b6e4056aa571/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 965.835998] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Uploaded image db3e6c6f-cc3d-4ae3-90f2-666514c372e9 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 965.839039] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 965.839572] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1427e011-3cc8-4a49-94c0-aa5a429a38b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.848654] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 965.848654] env[61986]: value = "task-1160127" [ 965.848654] env[61986]: _type = "Task" [ 965.848654] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.858182] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160127, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.042429] env[61986]: DEBUG nova.compute.manager [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Received event network-vif-deleted-ffee6438-b805-4a00-94a9-83fa6e39a230 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.042662] env[61986]: DEBUG nova.compute.manager [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.042831] env[61986]: DEBUG nova.compute.manager [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing instance network info cache due to event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 966.043060] env[61986]: DEBUG oslo_concurrency.lockutils [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.043207] env[61986]: DEBUG oslo_concurrency.lockutils [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.043377] env[61986]: DEBUG nova.network.neutron [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.083375] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.083658] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.083870] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.084092] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.084272] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.084451] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.084659] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.084832] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.085012] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.085356] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.085416] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.091761] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee203433-ae18-4c9c-bd95-66ce0844f82d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.117559] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 966.117559] env[61986]: value = "task-1160128" [ 966.117559] env[61986]: _type = "Task" [ 966.117559] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.125487] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160128, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.215886] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.223349] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160126, 'name': PowerOffVM_Task, 'duration_secs': 0.361842} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.223854] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.224721] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.225155] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1a98545-5500-439a-8ccf-14f7e8922528 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.231371] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160125, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628667} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.232694] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3b925174-cf0b-4c6c-b6b6-1c99a50026be/3b925174-cf0b-4c6c-b6b6-1c99a50026be.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 966.232694] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 966.233164] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af56fde2-533f-49d3-9846-02a80b14ff26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.241380] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 966.241380] env[61986]: value = "task-1160130" [ 966.241380] env[61986]: _type = "Task" [ 966.241380] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.255074] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.305373] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.559s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.309144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.544s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.310800] env[61986]: INFO nova.compute.claims [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.364417] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160127, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.632131] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160128, 'name': ReconfigVM_Task, 'duration_secs': 0.369849} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.632131] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 966.719856] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.759986] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155125} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.760469] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 966.761577] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1180ce60-aa37-44e6-b8c7-f665e3d092d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.799802] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 3b925174-cf0b-4c6c-b6b6-1c99a50026be/3b925174-cf0b-4c6c-b6b6-1c99a50026be.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.800861] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d45b4b6d-2eb8-466f-9757-03d8e2402ed4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.839104] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 966.839104] env[61986]: value = "task-1160131" [ 966.839104] env[61986]: _type = "Task" [ 966.839104] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.861641] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160131, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.866322] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160127, 'name': Destroy_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.896084] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.896736] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.896736] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Deleting the datastore file [datastore1] 021a27ec-cee0-454d-8daf-e6a82bd9330c {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.897027] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbb7b962-6e03-4cac-9af2-b5030ba41c19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.906222] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for the task: (returnval){ [ 966.906222] env[61986]: value = "task-1160132" [ 966.906222] env[61986]: _type = "Task" [ 966.906222] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.918375] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.921079] env[61986]: INFO nova.scheduler.client.report [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocation for migration b772f020-d14b-4ce6-a2c7-eb4d645079b2 [ 967.138179] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.138445] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.138667] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.138874] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.139036] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.139191] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.139402] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.139651] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.139875] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.140069] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.140369] env[61986]: DEBUG nova.virt.hardware [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.146283] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Reconfiguring VM instance instance-00000035 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 967.146597] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ceaa885-a281-4315-b2c3-59743c0afd49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.166757] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 967.166757] env[61986]: value = "task-1160133" [ 967.166757] env[61986]: _type = "Task" [ 967.166757] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.179658] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160133, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.350167] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160131, 'name': ReconfigVM_Task, 'duration_secs': 0.317303} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.350448] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 3b925174-cf0b-4c6c-b6b6-1c99a50026be/3b925174-cf0b-4c6c-b6b6-1c99a50026be.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.351257] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c45b8ca-e9d5-4134-bcfb-f2797a9a55e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.365531] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160127, 'name': Destroy_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.367226] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 967.367226] env[61986]: value = "task-1160134" [ 967.367226] env[61986]: _type = "Task" [ 967.367226] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.422398] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.437316] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d90a2bab-9591-4497-b7ed-f5480965cb56 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 17.474s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.440201] env[61986]: DEBUG nova.network.neutron [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updated VIF entry in instance network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.440624] env[61986]: DEBUG nova.network.neutron [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.681821] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.731679] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7841f980-5c0f-482e-94d0-74c378599c0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.740733] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3892b255-eabd-4aec-a771-26a211350441 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.776587] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c2cc3d-9474-489f-9860-c1ca68e482c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.786053] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80900885-4a0c-4817-a480-b2615f53f38d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.801178] env[61986]: DEBUG nova.compute.provider_tree [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.865030] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160127, 'name': Destroy_Task, 'duration_secs': 1.789034} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.865412] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Destroyed the VM [ 967.865752] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 967.866068] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-77606363-cb15-458f-adcb-0229e87291f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.881418] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160134, 'name': Rename_Task, 'duration_secs': 0.204391} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.883526] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 967.883526] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 967.883526] env[61986]: value = "task-1160135" [ 967.883526] env[61986]: _type = "Task" [ 967.883526] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.884157] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07e8ad78-4dcd-4d97-96b6-5e6e12806017 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.898715] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160135, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.901035] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 967.901035] env[61986]: value = "task-1160136" [ 967.901035] env[61986]: _type = "Task" [ 967.901035] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.910305] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.922726] env[61986]: DEBUG oslo_vmware.api [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Task: {'id': task-1160132, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.600654} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.922726] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.922726] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 967.922726] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 967.922726] env[61986]: INFO nova.compute.manager [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Took 2.25 seconds to destroy the instance on the hypervisor. [ 967.923272] env[61986]: DEBUG oslo.service.loopingcall [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.923770] env[61986]: DEBUG nova.compute.manager [-] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.923770] env[61986]: DEBUG nova.network.neutron [-] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 967.944985] env[61986]: DEBUG oslo_concurrency.lockutils [req-1d0af4dc-e90d-4ebe-a58b-78b00ebc5af0 req-3994a934-c518-490f-b299-04dfe099b3c9 service nova] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.181168] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160133, 'name': ReconfigVM_Task, 'duration_secs': 0.78681} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.181168] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Reconfigured VM instance instance-00000035 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 968.182293] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe2a0d0-5166-414a-88e9-a9aedf7d7a63 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.206436] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f/92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.207784] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3961402a-ff5e-4c05-b52c-4e65cfd0dc5d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.227387] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 968.227387] env[61986]: value = "task-1160137" [ 968.227387] env[61986]: _type = "Task" [ 968.227387] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.242851] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.290378] env[61986]: DEBUG nova.compute.manager [req-e781dcb2-b2ba-4feb-9476-cb0e092e8b7a req-3dfa52d3-858c-4075-81fe-404215fe0e32 service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Received event network-vif-deleted-9456eeae-0969-4fb6-b614-a699facd11f9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 968.290378] env[61986]: INFO nova.compute.manager [req-e781dcb2-b2ba-4feb-9476-cb0e092e8b7a req-3dfa52d3-858c-4075-81fe-404215fe0e32 service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Neutron deleted interface 9456eeae-0969-4fb6-b614-a699facd11f9; detaching it from the instance and deleting it from the info cache [ 968.290496] env[61986]: DEBUG nova.network.neutron [req-e781dcb2-b2ba-4feb-9476-cb0e092e8b7a req-3dfa52d3-858c-4075-81fe-404215fe0e32 service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.304754] env[61986]: DEBUG nova.scheduler.client.report [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.397681] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160135, 'name': RemoveSnapshot_Task} progress is 68%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.411627] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160136, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.742924] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.762237] env[61986]: DEBUG nova.network.neutron [-] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.793586] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b4d74fb-74bd-4e57-ab19-7824907f39fd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.806384] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fa3f53-236f-4729-a7e1-02bef121370d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.819660] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.820217] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 968.822897] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.356s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.827021] env[61986]: INFO nova.compute.claims [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 968.854189] env[61986]: DEBUG nova.compute.manager [req-e781dcb2-b2ba-4feb-9476-cb0e092e8b7a req-3dfa52d3-858c-4075-81fe-404215fe0e32 service nova] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Detach interface failed, port_id=9456eeae-0969-4fb6-b614-a699facd11f9, reason: Instance 021a27ec-cee0-454d-8daf-e6a82bd9330c could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 968.897359] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160135, 'name': RemoveSnapshot_Task} progress is 68%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.911135] env[61986]: DEBUG oslo_vmware.api [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160136, 'name': PowerOnVM_Task, 'duration_secs': 0.681218} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.911400] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 968.911604] env[61986]: INFO nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Took 8.72 seconds to spawn the instance on the hypervisor. [ 968.911787] env[61986]: DEBUG nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.912581] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b741f7-a101-4a0d-9b0f-e5e85aeb63b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.971389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.971669] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.971855] env[61986]: INFO nova.compute.manager [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Shelving [ 969.238611] env[61986]: DEBUG oslo_vmware.api [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160137, 'name': ReconfigVM_Task, 'duration_secs': 0.624324} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.238909] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f/92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.239198] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 50 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 969.267701] env[61986]: INFO nova.compute.manager [-] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Took 1.34 seconds to deallocate network for instance. [ 969.332289] env[61986]: DEBUG nova.compute.utils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 969.333759] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.333979] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.334191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.334373] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.334536] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.336110] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 969.336266] env[61986]: DEBUG nova.network.neutron [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 969.338327] env[61986]: INFO nova.compute.manager [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Terminating instance [ 969.342167] env[61986]: DEBUG nova.compute.manager [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.342167] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 969.342469] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0858524c-da16-4627-ba37-4c159b14f13b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.352211] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.352405] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba405729-7e8e-4ae3-92f0-b44a85999a76 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.360579] env[61986]: DEBUG oslo_vmware.api [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 969.360579] env[61986]: value = "task-1160138" [ 969.360579] env[61986]: _type = "Task" [ 969.360579] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.370370] env[61986]: DEBUG oslo_vmware.api [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.386276] env[61986]: DEBUG nova.policy [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 969.398044] env[61986]: DEBUG oslo_vmware.api [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160135, 'name': RemoveSnapshot_Task, 'duration_secs': 1.094265} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.398370] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 969.398643] env[61986]: INFO nova.compute.manager [None req-23c0122e-81c4-47d0-a6cd-db00a233995d tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Took 18.88 seconds to snapshot the instance on the hypervisor. [ 969.433625] env[61986]: INFO nova.compute.manager [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Took 28.87 seconds to build instance. [ 969.481909] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.482538] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d416dea5-3a4a-4696-a219-a4c562994084 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.491526] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 969.491526] env[61986]: value = "task-1160139" [ 969.491526] env[61986]: _type = "Task" [ 969.491526] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.504258] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.745749] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6bd7ce-13ea-45fe-8a64-a08573629400 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.780220] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.780220] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393de22b-909a-44fa-8412-fb14f1e7850e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.802919] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 67 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 969.820403] env[61986]: DEBUG nova.network.neutron [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Successfully created port: 74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 969.841120] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 969.874704] env[61986]: DEBUG oslo_vmware.api [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160138, 'name': PowerOffVM_Task, 'duration_secs': 0.281148} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.875011] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 969.875198] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 969.875455] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a38556d-6d14-4fee-99b1-b5a8e8ca7ba8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.905141] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.905434] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.935837] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c0f63557-71e8-4bb9-9441-988a84c58488 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.391s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.945727] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 969.945727] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 969.945921] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleting the datastore file [datastore1] f42e92ef-8a90-473c-8662-57cbb9e3f4ee {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.950021] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4168cc15-49d3-4a1c-a5c0-11806d2f6868 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.954163] env[61986]: DEBUG oslo_vmware.api [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 969.954163] env[61986]: value = "task-1160141" [ 969.954163] env[61986]: _type = "Task" [ 969.954163] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.966262] env[61986]: DEBUG oslo_vmware.api [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.004230] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160139, 'name': PowerOffVM_Task, 'duration_secs': 0.267425} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.004230] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 970.006998] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81dd0dce-8ea9-4da2-9174-28828e35ed08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.030073] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54e773d-d2c2-4ba7-a5fa-a56d6c711d66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.203229] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d54c14-19e2-45d6-a10d-328e60916208 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.211078] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e17991c-40ba-4bfd-9eee-af90685bf6ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.245623] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7376e881-4929-46cb-907c-0cc83751c88d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.255377] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058945d8-1cea-4399-be10-b0b8fcf4d5ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.271330] env[61986]: DEBUG nova.compute.provider_tree [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 970.350611] env[61986]: DEBUG nova.network.neutron [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Port 8bca3360-2dd5-4113-85c7-5e2e44864b07 binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 970.408635] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.467630] env[61986]: DEBUG oslo_vmware.api [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160141, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133734} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.467918] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.468122] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 970.468536] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 970.468536] env[61986]: INFO nova.compute.manager [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Took 1.13 seconds to destroy the instance on the hypervisor. [ 970.468729] env[61986]: DEBUG oslo.service.loopingcall [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.469639] env[61986]: DEBUG nova.compute.manager [-] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.469751] env[61986]: DEBUG nova.network.neutron [-] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 970.542323] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 970.542757] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-805458a6-084f-4250-ad33-57acedb66825 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.568811] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 970.568811] env[61986]: value = "task-1160142" [ 970.568811] env[61986]: _type = "Task" [ 970.568811] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.579012] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160142, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.622098] env[61986]: DEBUG nova.compute.manager [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.623025] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd36d91a-8d2e-4f53-8553-d09cd7d9c204 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.757386] env[61986]: DEBUG nova.compute.manager [req-b2aca0cf-f404-4a6c-880d-b3c8997fda2d req-7066d6c5-6cb1-4a79-8f42-2ea27138250e service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Received event network-vif-deleted-e7c53e1f-f6f4-4df9-b4b0-e2818f87b045 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.757504] env[61986]: INFO nova.compute.manager [req-b2aca0cf-f404-4a6c-880d-b3c8997fda2d req-7066d6c5-6cb1-4a79-8f42-2ea27138250e service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Neutron deleted interface e7c53e1f-f6f4-4df9-b4b0-e2818f87b045; detaching it from the instance and deleting it from the info cache [ 970.757696] env[61986]: DEBUG nova.network.neutron [req-b2aca0cf-f404-4a6c-880d-b3c8997fda2d req-7066d6c5-6cb1-4a79-8f42-2ea27138250e service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.795416] env[61986]: ERROR nova.scheduler.client.report [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [req-1160f8f8-259f-4db0-8b37-bd38668f2fc2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1160f8f8-259f-4db0-8b37-bd38668f2fc2"}]} [ 970.819931] env[61986]: DEBUG nova.scheduler.client.report [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Refreshing inventories for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 970.836609] env[61986]: DEBUG nova.scheduler.client.report [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updating ProviderTree inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 970.836851] env[61986]: DEBUG nova.compute.provider_tree [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 970.849061] env[61986]: DEBUG nova.scheduler.client.report [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Refreshing aggregate associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, aggregates: None {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 970.853450] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 970.870821] env[61986]: DEBUG nova.scheduler.client.report [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Refreshing trait associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 970.875916] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.876192] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.876360] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.876568] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.876721] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.876874] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.877102] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.877286] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.877485] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.877775] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.878015] env[61986]: DEBUG nova.virt.hardware [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.878972] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f84e10-a27d-4755-8fa5-c5c79653e1f8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.888324] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66c8eed-e180-438f-9a96-a871e220ebd4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.927538] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.079336] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160142, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.139464] env[61986]: INFO nova.compute.manager [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] instance snapshotting [ 971.144948] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946b94a6-63ce-4e38-ab55-e1b2604f6f0d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.173560] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4344065-994c-4b58-96fe-1c0797141253 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.188154] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563e8723-e404-4c87-a787-29083f4bd9bd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.196927] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734236b6-6b01-4b3a-9b37-7a09c6e18436 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.230360] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44041041-5c55-474c-8eaf-02b1830159da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.235838] env[61986]: DEBUG nova.network.neutron [-] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.241561] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb0ff4d-cc20-4e95-b894-4a4bd641b214 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.257169] env[61986]: DEBUG nova.compute.provider_tree [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.260751] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-95e69055-42e9-40de-aa22-9531714ecafe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.271569] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7abe47e-55bb-49bc-b7b9-682351d08130 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.307969] env[61986]: DEBUG nova.compute.manager [req-b2aca0cf-f404-4a6c-880d-b3c8997fda2d req-7066d6c5-6cb1-4a79-8f42-2ea27138250e service nova] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Detach interface failed, port_id=e7c53e1f-f6f4-4df9-b4b0-e2818f87b045, reason: Instance f42e92ef-8a90-473c-8662-57cbb9e3f4ee could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 971.376267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.376522] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.376705] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.487270] env[61986]: DEBUG nova.compute.manager [req-63af2e2f-fa5b-4364-a9f7-a093d62c0720 req-0ef51935-7e16-48d5-ae8e-1154079c6c73 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-vif-plugged-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.487549] env[61986]: DEBUG oslo_concurrency.lockutils [req-63af2e2f-fa5b-4364-a9f7-a093d62c0720 req-0ef51935-7e16-48d5-ae8e-1154079c6c73 service nova] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.487816] env[61986]: DEBUG oslo_concurrency.lockutils [req-63af2e2f-fa5b-4364-a9f7-a093d62c0720 req-0ef51935-7e16-48d5-ae8e-1154079c6c73 service nova] Lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.488353] env[61986]: DEBUG oslo_concurrency.lockutils [req-63af2e2f-fa5b-4364-a9f7-a093d62c0720 req-0ef51935-7e16-48d5-ae8e-1154079c6c73 service nova] Lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.488664] env[61986]: DEBUG nova.compute.manager [req-63af2e2f-fa5b-4364-a9f7-a093d62c0720 req-0ef51935-7e16-48d5-ae8e-1154079c6c73 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] No waiting events found dispatching network-vif-plugged-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 971.488978] env[61986]: WARNING nova.compute.manager [req-63af2e2f-fa5b-4364-a9f7-a093d62c0720 req-0ef51935-7e16-48d5-ae8e-1154079c6c73 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received unexpected event network-vif-plugged-74f613a6-57a3-4f00-b018-8c58f254200c for instance with vm_state building and task_state spawning. [ 971.578039] env[61986]: DEBUG nova.network.neutron [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Successfully updated port: 74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.582986] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160142, 'name': CreateSnapshot_Task, 'duration_secs': 0.841894} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.583066] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 971.583786] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746bb836-d948-4722-9661-d951cfe4321f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.646085] env[61986]: DEBUG nova.compute.manager [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.647227] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc2350e-91b8-4df5-b79c-e0e229950d66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.692147] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 971.692744] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0f3ce5a9-4cbf-4a9b-aab4-37d7b91d2c8a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.702861] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 971.702861] env[61986]: value = "task-1160143" [ 971.702861] env[61986]: _type = "Task" [ 971.702861] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.712144] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160143, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.738326] env[61986]: INFO nova.compute.manager [-] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Took 1.27 seconds to deallocate network for instance. [ 971.796414] env[61986]: DEBUG nova.scheduler.client.report [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 971.796717] env[61986]: DEBUG nova.compute.provider_tree [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 107 to 108 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 971.796927] env[61986]: DEBUG nova.compute.provider_tree [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.084522] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.084766] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.084864] env[61986]: DEBUG nova.network.neutron [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 972.104484] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 972.104966] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e416bf36-b23c-42e7-8c80-9046feb24bfa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.117530] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 972.117530] env[61986]: value = "task-1160144" [ 972.117530] env[61986]: _type = "Task" [ 972.117530] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.126559] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160144, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.158686] env[61986]: INFO nova.compute.manager [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] instance snapshotting [ 972.161703] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84ab6a8-e36f-461d-a09e-39b3ef6e6422 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.181410] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259c5faf-a820-4406-8d5d-fef5206f4751 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.212845] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160143, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.245633] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.302160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.479s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.302754] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 972.305431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.897s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.306873] env[61986]: INFO nova.compute.claims [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.451237] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.451627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.451973] env[61986]: DEBUG nova.network.neutron [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 972.619416] env[61986]: DEBUG nova.network.neutron [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 972.630833] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160144, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.694638] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 972.695116] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-05acdedb-cc24-44ee-a6ad-8359127d5b5a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.705764] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 972.705764] env[61986]: value = "task-1160145" [ 972.705764] env[61986]: _type = "Task" [ 972.705764] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.719279] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160143, 'name': CreateSnapshot_Task, 'duration_secs': 0.841804} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.721552] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 972.721921] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160145, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.722552] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8323d4b2-31f3-4064-8ec5-b880d0ecb8b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.783904] env[61986]: DEBUG nova.network.neutron [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.812164] env[61986]: DEBUG nova.compute.utils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.815839] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 972.816055] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 972.862290] env[61986]: DEBUG nova.policy [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8be836eb63c544e4ab5a831b3397a898', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c64e7f5b4b64d1988edf2a2fcbe32eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 973.132700] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160144, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.210984] env[61986]: DEBUG nova.network.neutron [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.223735] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160145, 'name': CreateSnapshot_Task, 'duration_secs': 0.47474} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.224035] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 973.225162] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add5f2e8-aff7-4105-bf75-c063aa57506d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.231875] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Successfully created port: 759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.243424] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 973.246109] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9d3dc837-0e9d-4937-ba85-44480f69bd2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.255160] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 973.255160] env[61986]: value = "task-1160146" [ 973.255160] env[61986]: _type = "Task" [ 973.255160] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.265100] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160146, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.286530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.286867] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Instance network_info: |[{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 973.287394] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:c9:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e0cfc48-d93b-4477-8082-69a2f7aa7701', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74f613a6-57a3-4f00-b018-8c58f254200c', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.294938] env[61986]: DEBUG oslo.service.loopingcall [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.295440] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 973.295675] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc5b09fe-bb9e-4906-a9d5-c85a89d52214 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.316353] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 973.321526] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.321526] env[61986]: value = "task-1160147" [ 973.321526] env[61986]: _type = "Task" [ 973.321526] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.331140] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160147, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.548903] env[61986]: DEBUG nova.compute.manager [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 973.548903] env[61986]: DEBUG nova.compute.manager [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing instance network info cache due to event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 973.548903] env[61986]: DEBUG oslo_concurrency.lockutils [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.548903] env[61986]: DEBUG oslo_concurrency.lockutils [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.548903] env[61986]: DEBUG nova.network.neutron [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 973.630673] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160144, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.650749] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Successfully created port: b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.694899] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046857b7-829d-4190-887f-56c4b17a43d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.703659] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1517fe22-a834-4b22-a980-79affbceed05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.741648] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.748045] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e678d3-810a-4141-8758-a27c5ef69e70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.759699] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 973.760526] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7b4aa29b-ecac-4efa-935c-0bdb3605046d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.773885] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc313716-d58f-43db-8263-515a2df2e398 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.784984] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160146, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.785812] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 973.785812] env[61986]: value = "task-1160148" [ 973.785812] env[61986]: _type = "Task" [ 973.785812] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.799701] env[61986]: DEBUG nova.compute.provider_tree [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.808956] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160148, 'name': CloneVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.836337] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160147, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.965066] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Successfully created port: 0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.131692] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160144, 'name': CloneVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.273073] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160146, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.274219] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1844bfe-2b45-470d-934c-cdce35500f1d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.301757] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a5852d-c09f-4c99-a681-0f5a8b79e4eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.313494] env[61986]: DEBUG nova.scheduler.client.report [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.317994] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 83 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 974.325451] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160148, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.330835] env[61986]: DEBUG nova.network.neutron [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updated VIF entry in instance network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 974.331192] env[61986]: DEBUG nova.network.neutron [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.333043] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 974.342665] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160147, 'name': CreateVM_Task, 'duration_secs': 0.970912} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.342828] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 974.343535] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.343700] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.344104] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.344875] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb117094-0f9c-4b42-ad83-18ca9ca63d79 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.351323] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 974.351323] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528d0224-1a29-0d7a-05c3-46a52cbe701b" [ 974.351323] env[61986]: _type = "Task" [ 974.351323] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.362696] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528d0224-1a29-0d7a-05c3-46a52cbe701b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.365070] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.365326] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.365486] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.365671] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.369017] env[61986]: DEBUG nova.virt.hardware [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.369017] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1c450c-00e3-4e7c-bc9c-05439b94152e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.376415] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232adc12-318d-4065-8196-3d2c7ae81ad3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.631411] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160144, 'name': CloneVM_Task, 'duration_secs': 2.066194} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.631692] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Created linked-clone VM from snapshot [ 974.632460] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733e3738-d802-4277-ba7d-341c20d34daf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.640267] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Uploading image 24ded09f-c9f0-435c-93fc-1aef721bbca3 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 974.665495] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 974.665495] env[61986]: value = "vm-252418" [ 974.665495] env[61986]: _type = "VirtualMachine" [ 974.665495] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 974.665887] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4a6c393e-6505-4c7d-a7ef-47b15f3f681b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.675716] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lease: (returnval){ [ 974.675716] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed56d3-9d22-1d28-77fc-75cc38f3d059" [ 974.675716] env[61986]: _type = "HttpNfcLease" [ 974.675716] env[61986]: } obtained for exporting VM: (result){ [ 974.675716] env[61986]: value = "vm-252418" [ 974.675716] env[61986]: _type = "VirtualMachine" [ 974.675716] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 974.676034] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the lease: (returnval){ [ 974.676034] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed56d3-9d22-1d28-77fc-75cc38f3d059" [ 974.676034] env[61986]: _type = "HttpNfcLease" [ 974.676034] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 974.685168] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 974.685168] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed56d3-9d22-1d28-77fc-75cc38f3d059" [ 974.685168] env[61986]: _type = "HttpNfcLease" [ 974.685168] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 974.771648] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160146, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.809349] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160148, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.826460] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.827074] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.831576] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f4e32d0-b4ed-4d5b-9c74-0d1b498c4e11 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance '92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f' progress to 100 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 974.835316] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.410s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.835502] env[61986]: DEBUG nova.objects.instance [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 974.838691] env[61986]: DEBUG oslo_concurrency.lockutils [req-90077536-1175-4560-a600-a71cee774f76 req-cd790ba5-07a8-4bf7-97ad-4ceea989e9b6 service nova] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.862618] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528d0224-1a29-0d7a-05c3-46a52cbe701b, 'name': SearchDatastore_Task, 'duration_secs': 0.012291} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.862954] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.863285] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.863500] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.863911] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.863911] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.864146] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c083b009-6e54-4d9c-b658-0d1aa94b8f71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.874100] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.874328] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 974.875174] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-213ce100-9a53-46b6-a638-8dab08862d7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.885170] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 974.885170] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52057ef5-2bb8-5a0b-9cb4-f9a5906149ff" [ 974.885170] env[61986]: _type = "Task" [ 974.885170] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.897534] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52057ef5-2bb8-5a0b-9cb4-f9a5906149ff, 'name': SearchDatastore_Task, 'duration_secs': 0.010065} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.898697] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2219fbfd-c981-442f-bf2e-050432073433 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.907912] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 974.907912] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5292922e-cf21-f624-0719-f63ec622bcc6" [ 974.907912] env[61986]: _type = "Task" [ 974.907912] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.919554] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5292922e-cf21-f624-0719-f63ec622bcc6, 'name': SearchDatastore_Task, 'duration_secs': 0.0097} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.919938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.920332] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] bda7750f-eec9-40d9-ace5-18d48234126e/bda7750f-eec9-40d9-ace5-18d48234126e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 974.920644] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c588884-88a9-4605-af7c-c75b0295d2d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.931322] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 974.931322] env[61986]: value = "task-1160150" [ 974.931322] env[61986]: _type = "Task" [ 974.931322] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.940585] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160150, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.187382] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 975.187382] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed56d3-9d22-1d28-77fc-75cc38f3d059" [ 975.187382] env[61986]: _type = "HttpNfcLease" [ 975.187382] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 975.188408] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 975.188408] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ed56d3-9d22-1d28-77fc-75cc38f3d059" [ 975.188408] env[61986]: _type = "HttpNfcLease" [ 975.188408] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 975.189261] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1445b7c-4cb6-4b5d-b814-371bbad3772c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.201855] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523df22d-8f8f-838b-1b63-004eb9ab19ea/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 975.202087] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523df22d-8f8f-838b-1b63-004eb9ab19ea/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 975.303665] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160146, 'name': CloneVM_Task, 'duration_secs': 1.824408} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.308756] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Created linked-clone VM from snapshot [ 975.310120] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a609ae-078f-4b24-8b87-9659c66fa217 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.323212] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Uploading image 9778a241-91dc-4636-b2c0-5e8324c381e4 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 975.331352] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160148, 'name': CloneVM_Task, 'duration_secs': 1.323204} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.331774] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Created linked-clone VM from snapshot [ 975.332943] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220c997e-636b-459c-9f27-05a1c5fc0f94 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.344124] env[61986]: DEBUG nova.compute.utils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.353912] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 975.358122] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.358468] env[61986]: DEBUG nova.network.neutron [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 975.361710] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Uploading image 7d7a5833-c7c3-4049-9eab-4941b2fe99d6 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 975.364239] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5a75cc91-7795-4fb1-a910-f75f0dfb270d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.369600] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.381577] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 975.381577] env[61986]: value = "task-1160151" [ 975.381577] env[61986]: _type = "Task" [ 975.381577] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.394329] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ae58c573-14a5-4dad-981e-c29e1ebdd517 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.402499] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160151, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.407543] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 975.407543] env[61986]: value = "vm-252423" [ 975.407543] env[61986]: _type = "VirtualMachine" [ 975.407543] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 975.408243] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e84c1e2b-677d-4035-860e-9934df8e0954 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.423715] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lease: (returnval){ [ 975.423715] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbe407-a087-5415-e16d-f26cf505f53b" [ 975.423715] env[61986]: _type = "HttpNfcLease" [ 975.423715] env[61986]: } obtained for exporting VM: (result){ [ 975.423715] env[61986]: value = "vm-252423" [ 975.423715] env[61986]: _type = "VirtualMachine" [ 975.423715] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 975.424043] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the lease: (returnval){ [ 975.424043] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbe407-a087-5415-e16d-f26cf505f53b" [ 975.424043] env[61986]: _type = "HttpNfcLease" [ 975.424043] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 975.435755] env[61986]: DEBUG nova.policy [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e34c464681464270a5ae8c02b6e3fb75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ea5fc77300041f5aeccb8773300ef0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 975.439777] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 975.439777] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbe407-a087-5415-e16d-f26cf505f53b" [ 975.439777] env[61986]: _type = "HttpNfcLease" [ 975.439777] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 975.445974] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160150, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46857} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.446249] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] bda7750f-eec9-40d9-ace5-18d48234126e/bda7750f-eec9-40d9-ace5-18d48234126e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 975.446469] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.446711] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a10741b3-6b28-4d82-a432-2ceb4c530376 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.456443] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 975.456443] env[61986]: value = "task-1160153" [ 975.456443] env[61986]: _type = "Task" [ 975.456443] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.465806] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.671059] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquiring lock "5170aa51-3307-42b1-b0dd-645dd4036e5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.675024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.675024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquiring lock "5170aa51-3307-42b1-b0dd-645dd4036e5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.675024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.675024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.677286] env[61986]: INFO nova.compute.manager [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Terminating instance [ 975.680961] env[61986]: DEBUG nova.compute.manager [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.681079] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 975.681362] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-198f2ad8-eb99-4e46-98d3-4ac17dcf7301 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.689290] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 975.689290] env[61986]: value = "task-1160154" [ 975.689290] env[61986]: _type = "Task" [ 975.689290] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.707302] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.735980] env[61986]: DEBUG nova.network.neutron [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Successfully created port: dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.868696] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cd19cd71-e1c4-4b02-9815-03dd4c5dfef6 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.033s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.870123] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.496s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.872388] env[61986]: INFO nova.compute.claims [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.895621] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160151, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.938593] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 975.938593] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbe407-a087-5415-e16d-f26cf505f53b" [ 975.938593] env[61986]: _type = "HttpNfcLease" [ 975.938593] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 975.939554] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 975.939554] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbe407-a087-5415-e16d-f26cf505f53b" [ 975.939554] env[61986]: _type = "HttpNfcLease" [ 975.939554] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 975.941624] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd33e90-5b5b-4be8-a045-0910e4ac2bc6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.953847] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254cc62-8e04-d799-2237-cc0523a3ac31/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 975.953847] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254cc62-8e04-d799-2237-cc0523a3ac31/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 976.038542] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06357} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.039103] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.040122] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfad450-abf0-479e-be0e-515cc8cc4cda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.066266] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] bda7750f-eec9-40d9-ace5-18d48234126e/bda7750f-eec9-40d9-ace5-18d48234126e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.067477] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9779ffa-5de9-4e1a-8268-9a20d594744f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.089792] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 976.089792] env[61986]: value = "task-1160155" [ 976.089792] env[61986]: _type = "Task" [ 976.089792] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.098445] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d06f0cf8-d016-44c4-be85-34b0977fc384 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.104306] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160155, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.108441] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Successfully updated port: 759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.162816] env[61986]: DEBUG nova.compute.manager [req-75fcf1cc-ed52-443b-acbe-8a06d79d9c71 req-8ddbe859-dbb4-4828-a28d-73cd74e83723 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-vif-plugged-759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.163068] env[61986]: DEBUG oslo_concurrency.lockutils [req-75fcf1cc-ed52-443b-acbe-8a06d79d9c71 req-8ddbe859-dbb4-4828-a28d-73cd74e83723 service nova] Acquiring lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.163280] env[61986]: DEBUG oslo_concurrency.lockutils [req-75fcf1cc-ed52-443b-acbe-8a06d79d9c71 req-8ddbe859-dbb4-4828-a28d-73cd74e83723 service nova] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.163443] env[61986]: DEBUG oslo_concurrency.lockutils [req-75fcf1cc-ed52-443b-acbe-8a06d79d9c71 req-8ddbe859-dbb4-4828-a28d-73cd74e83723 service nova] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.163612] env[61986]: DEBUG nova.compute.manager [req-75fcf1cc-ed52-443b-acbe-8a06d79d9c71 req-8ddbe859-dbb4-4828-a28d-73cd74e83723 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] No waiting events found dispatching network-vif-plugged-759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.163791] env[61986]: WARNING nova.compute.manager [req-75fcf1cc-ed52-443b-acbe-8a06d79d9c71 req-8ddbe859-dbb4-4828-a28d-73cd74e83723 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received unexpected event network-vif-plugged-759ea478-47fb-45dc-a788-f500f86de7fd for instance with vm_state building and task_state spawning. [ 976.201550] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160154, 'name': PowerOffVM_Task, 'duration_secs': 0.482476} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.202315] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 976.202526] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 976.202735] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252362', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'name': 'volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5170aa51-3307-42b1-b0dd-645dd4036e5b', 'attached_at': '', 'detached_at': '', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'serial': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 976.203842] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501326f9-3a0c-44aa-af19-d957ca56f845 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.225577] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52376c1a-e51e-4868-843c-0c302ce087bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.234208] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d94fe5b-31bc-4ac5-9780-c01baae6c2ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.254233] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ab5ba8-dc28-42c3-b4f9-987c0e33fc5f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.270517] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] The volume has not been displaced from its original location: [datastore1] volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5/volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 976.276994] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 976.277739] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-222ebbc4-c7bd-4114-b9ce-90d44741f434 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.299286] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 976.299286] env[61986]: value = "task-1160156" [ 976.299286] env[61986]: _type = "Task" [ 976.299286] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.309249] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.384534] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.405639] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160151, 'name': Destroy_Task, 'duration_secs': 0.634748} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.408864] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Destroyed the VM [ 976.409193] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 976.410666] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6d2cfbf4-9ce8-4677-b920-b4aa779c7e4c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.418236] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.418648] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.419068] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.419210] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.419437] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.419726] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.420010] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.420261] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.420534] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.421210] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.421288] env[61986]: DEBUG nova.virt.hardware [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.422980] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9d17e1-55b7-43db-94df-e7cdd1e93419 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.429387] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 976.429387] env[61986]: value = "task-1160157" [ 976.429387] env[61986]: _type = "Task" [ 976.429387] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.440048] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fded556e-5cf9-4dd6-9088-d69d40e85c1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.448342] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160157, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.603107] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160155, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.813829] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160156, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.945088] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160157, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.108182] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160155, 'name': ReconfigVM_Task, 'duration_secs': 0.533097} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.111724] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Reconfigured VM instance instance-00000054 to attach disk [datastore1] bda7750f-eec9-40d9-ace5-18d48234126e/bda7750f-eec9-40d9-ace5-18d48234126e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.113529] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-954c893d-76d4-45d6-ab1d-1c006dfa0755 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.124802] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 977.124802] env[61986]: value = "task-1160158" [ 977.124802] env[61986]: _type = "Task" [ 977.124802] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.135068] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160158, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.310805] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224ddc56-cbba-4c37-aba0-5c39533203f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.319591] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160156, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.326322] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b111091f-d577-45d9-b483-ea0afd0943d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.365991] env[61986]: DEBUG nova.network.neutron [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Successfully updated port: dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.369915] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1299b28-6425-40a8-b317-beae32683a49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.374932] env[61986]: DEBUG nova.compute.manager [req-13abe92c-4b5d-4832-8839-071ac1639136 req-2299ad5d-3048-40ab-80b5-863e47932661 service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Received event network-vif-plugged-dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.375334] env[61986]: DEBUG oslo_concurrency.lockutils [req-13abe92c-4b5d-4832-8839-071ac1639136 req-2299ad5d-3048-40ab-80b5-863e47932661 service nova] Acquiring lock "73b2c463-0460-4a68-b788-ab0c7818d8f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.375959] env[61986]: DEBUG oslo_concurrency.lockutils [req-13abe92c-4b5d-4832-8839-071ac1639136 req-2299ad5d-3048-40ab-80b5-863e47932661 service nova] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.376709] env[61986]: DEBUG oslo_concurrency.lockutils [req-13abe92c-4b5d-4832-8839-071ac1639136 req-2299ad5d-3048-40ab-80b5-863e47932661 service nova] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.376976] env[61986]: DEBUG nova.compute.manager [req-13abe92c-4b5d-4832-8839-071ac1639136 req-2299ad5d-3048-40ab-80b5-863e47932661 service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] No waiting events found dispatching network-vif-plugged-dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.377286] env[61986]: WARNING nova.compute.manager [req-13abe92c-4b5d-4832-8839-071ac1639136 req-2299ad5d-3048-40ab-80b5-863e47932661 service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Received unexpected event network-vif-plugged-dd370310-3b48-4932-8a9f-12a6e4298161 for instance with vm_state building and task_state spawning. [ 977.386659] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4df03b-83f7-44b1-8f00-1bcf25032ce0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.406609] env[61986]: DEBUG nova.compute.provider_tree [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.412057] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.412337] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.412564] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.413018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.413588] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.416105] env[61986]: INFO nova.compute.manager [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Terminating instance [ 977.418553] env[61986]: DEBUG nova.compute.manager [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.418990] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 977.422752] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c2d457-1c38-4a42-8ba4-3e7afb282d58 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.440036] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.441408] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63160518-f0fe-4032-b086-29854cd302ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.448038] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160157, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.450595] env[61986]: DEBUG oslo_vmware.api [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 977.450595] env[61986]: value = "task-1160159" [ 977.450595] env[61986]: _type = "Task" [ 977.450595] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.461916] env[61986]: DEBUG oslo_vmware.api [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.637473] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160158, 'name': Rename_Task, 'duration_secs': 0.20968} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.637849] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 977.638174] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f989aa1-df62-4f41-bd74-d3bb3baacff3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.647665] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 977.647665] env[61986]: value = "task-1160160" [ 977.647665] env[61986]: _type = "Task" [ 977.647665] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.659543] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160160, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.812296] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160156, 'name': ReconfigVM_Task, 'duration_secs': 1.286912} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.813034] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 977.818972] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25f7b618-373e-46b8-8846-12b00b16187b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.838099] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 977.838099] env[61986]: value = "task-1160161" [ 977.838099] env[61986]: _type = "Task" [ 977.838099] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.849530] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.869180] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "refresh_cache-73b2c463-0460-4a68-b788-ab0c7818d8f8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.869360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "refresh_cache-73b2c463-0460-4a68-b788-ab0c7818d8f8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.869521] env[61986]: DEBUG nova.network.neutron [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.933190] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.933552] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.933857] env[61986]: DEBUG nova.compute.manager [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Going to confirm migration 3 {{(pid=61986) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 977.948603] env[61986]: DEBUG oslo_vmware.api [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160157, 'name': RemoveSnapshot_Task, 'duration_secs': 1.412847} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.949042] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 977.953095] env[61986]: DEBUG nova.scheduler.client.report [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 108 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 977.953412] env[61986]: DEBUG nova.compute.provider_tree [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 108 to 109 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 977.953953] env[61986]: DEBUG nova.compute.provider_tree [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.971029] env[61986]: DEBUG oslo_vmware.api [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160159, 'name': PowerOffVM_Task, 'duration_secs': 0.38879} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.971421] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.971766] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.972203] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ddf13d29-701c-4220-96da-679860ea48bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.058636] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 978.058636] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 978.060077] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleting the datastore file [datastore1] f15bf19d-b86a-4b0a-ac1a-9df8e77e9382 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.060682] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5beaf28-1b73-42f4-a9f2-0e3319f42877 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.071130] env[61986]: DEBUG oslo_vmware.api [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 978.071130] env[61986]: value = "task-1160163" [ 978.071130] env[61986]: _type = "Task" [ 978.071130] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.089151] env[61986]: DEBUG oslo_vmware.api [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160163, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.159403] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160160, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.350034] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160161, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.416685] env[61986]: DEBUG nova.network.neutron [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.438030] env[61986]: DEBUG nova.compute.manager [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-changed-759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.438348] env[61986]: DEBUG nova.compute.manager [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Refreshing instance network info cache due to event network-changed-759ea478-47fb-45dc-a788-f500f86de7fd. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 978.438654] env[61986]: DEBUG oslo_concurrency.lockutils [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] Acquiring lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.438813] env[61986]: DEBUG oslo_concurrency.lockutils [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] Acquired lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.438976] env[61986]: DEBUG nova.network.neutron [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Refreshing network info cache for port 759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.460368] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.461099] env[61986]: DEBUG nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.465097] env[61986]: WARNING nova.compute.manager [None req-870078db-2d92-4137-b463-d759a50e31ea tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Image not found during snapshot: nova.exception.ImageNotFound: Image 9778a241-91dc-4636-b2c0-5e8324c381e4 could not be found. [ 978.466637] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.190s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.468094] env[61986]: INFO nova.compute.claims [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.571455] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.571682] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.571880] env[61986]: DEBUG nova.network.neutron [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.572096] env[61986]: DEBUG nova.objects.instance [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'info_cache' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.585576] env[61986]: DEBUG oslo_vmware.api [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160163, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.250824} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.586633] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.587074] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 978.587550] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 978.588070] env[61986]: INFO nova.compute.manager [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Took 1.17 seconds to destroy the instance on the hypervisor. [ 978.588357] env[61986]: DEBUG oslo.service.loopingcall [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.588579] env[61986]: DEBUG nova.compute.manager [-] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.588681] env[61986]: DEBUG nova.network.neutron [-] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 978.601413] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Successfully updated port: b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 978.661915] env[61986]: DEBUG oslo_vmware.api [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160160, 'name': PowerOnVM_Task, 'duration_secs': 0.694723} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.662092] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 978.663052] env[61986]: INFO nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Took 7.81 seconds to spawn the instance on the hypervisor. [ 978.663052] env[61986]: DEBUG nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.663461] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecba669-bd6c-4dfe-99d3-eda4f9303702 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.701629] env[61986]: DEBUG nova.network.neutron [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Updating instance_info_cache with network_info: [{"id": "dd370310-3b48-4932-8a9f-12a6e4298161", "address": "fa:16:3e:18:01:e3", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd370310-3b", "ovs_interfaceid": "dd370310-3b48-4932-8a9f-12a6e4298161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.851246] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160161, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.969282] env[61986]: DEBUG nova.compute.utils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.970920] env[61986]: DEBUG nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 979.016180] env[61986]: DEBUG nova.network.neutron [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 979.166101] env[61986]: DEBUG nova.network.neutron [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.183314] env[61986]: INFO nova.compute.manager [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Took 24.44 seconds to build instance. [ 979.207073] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "refresh_cache-73b2c463-0460-4a68-b788-ab0c7818d8f8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.207073] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Instance network_info: |[{"id": "dd370310-3b48-4932-8a9f-12a6e4298161", "address": "fa:16:3e:18:01:e3", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd370310-3b", "ovs_interfaceid": "dd370310-3b48-4932-8a9f-12a6e4298161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.207073] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:01:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0954fad3-d24d-496c-83e6-a09d3cb556fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd370310-3b48-4932-8a9f-12a6e4298161', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.213891] env[61986]: DEBUG oslo.service.loopingcall [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.214815] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 979.215258] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ff54ad1-de9b-4e67-8551-d7531f017f7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.244426] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.244426] env[61986]: value = "task-1160164" [ 979.244426] env[61986]: _type = "Task" [ 979.244426] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.254945] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160164, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.353117] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160161, 'name': ReconfigVM_Task, 'duration_secs': 1.169107} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.353618] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252362', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'name': 'volume-5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5170aa51-3307-42b1-b0dd-645dd4036e5b', 'attached_at': '', 'detached_at': '', 'volume_id': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5', 'serial': '5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 979.353955] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 979.354829] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c13d3af-3e1a-482e-8c28-1d5ab77dc3f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.364743] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 979.365169] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f076d6a4-4b5b-481a-b7a0-e93118736c30 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.455524] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.455940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.456226] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Deleting the datastore file [datastore1] 5170aa51-3307-42b1-b0dd-645dd4036e5b {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.456651] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af8d23f0-e229-4b46-8141-91381ee911b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.466051] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for the task: (returnval){ [ 979.466051] env[61986]: value = "task-1160166" [ 979.466051] env[61986]: _type = "Task" [ 979.466051] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.476923] env[61986]: DEBUG nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.478964] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.668786] env[61986]: DEBUG oslo_concurrency.lockutils [req-49779174-64f5-4475-af51-629b235dc268 req-fa6b8984-5791-4ec1-9776-c957bcad2d9a service nova] Releasing lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.688949] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d5ab829-e734-43a8-8767-ddd8369be4ac tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bda7750f-eec9-40d9-ace5-18d48234126e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.955s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.760390] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160164, 'name': CreateVM_Task, 'duration_secs': 0.404215} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.760577] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 979.762657] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.762657] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.763684] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.764309] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-624fcc4b-92a5-49cc-8a05-1eee2cfa6a90 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.769637] env[61986]: DEBUG nova.network.neutron [-] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.776635] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 979.776635] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cc033d-74e9-7300-4984-b19cde461f17" [ 979.776635] env[61986]: _type = "Task" [ 979.776635] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.787493] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cc033d-74e9-7300-4984-b19cde461f17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.875185] env[61986]: DEBUG nova.network.neutron [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.910852] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f90bbc-2e69-426e-88c2-896291b9245f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.919464] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add6534b-3bdc-4a39-9c24-033ebc256549 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.953634] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54733644-27d1-434b-bcd9-bf107ff6f928 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.963346] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c10a66-0122-4629-a865-848f0a6bbddc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.985393] env[61986]: DEBUG nova.compute.provider_tree [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.989780] env[61986]: DEBUG oslo_vmware.api [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Task: {'id': task-1160166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166303} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.990560] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.991940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 979.991940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 979.991940] env[61986]: INFO nova.compute.manager [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Took 4.31 seconds to destroy the instance on the hypervisor. [ 979.991940] env[61986]: DEBUG oslo.service.loopingcall [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.991940] env[61986]: DEBUG nova.compute.manager [-] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.991940] env[61986]: DEBUG nova.network.neutron [-] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.023417] env[61986]: DEBUG nova.compute.manager [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Received event network-changed-dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.023734] env[61986]: DEBUG nova.compute.manager [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Refreshing instance network info cache due to event network-changed-dd370310-3b48-4932-8a9f-12a6e4298161. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 980.023958] env[61986]: DEBUG oslo_concurrency.lockutils [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] Acquiring lock "refresh_cache-73b2c463-0460-4a68-b788-ab0c7818d8f8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.024038] env[61986]: DEBUG oslo_concurrency.lockutils [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] Acquired lock "refresh_cache-73b2c463-0460-4a68-b788-ab0c7818d8f8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.024291] env[61986]: DEBUG nova.network.neutron [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Refreshing network info cache for port dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 980.180827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.180827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.180827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.181137] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.181137] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.184059] env[61986]: INFO nova.compute.manager [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Terminating instance [ 980.185404] env[61986]: DEBUG nova.compute.manager [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.185686] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 980.186843] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64274cc-d136-468f-9b57-d0d5537f4795 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.195942] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 980.196271] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e9539d9-5f74-43af-bd0f-eefc61bfbd4a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.204733] env[61986]: DEBUG oslo_vmware.api [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 980.204733] env[61986]: value = "task-1160167" [ 980.204733] env[61986]: _type = "Task" [ 980.204733] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.214720] env[61986]: DEBUG oslo_vmware.api [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.276733] env[61986]: INFO nova.compute.manager [-] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Took 1.69 seconds to deallocate network for instance. [ 980.295953] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cc033d-74e9-7300-4984-b19cde461f17, 'name': SearchDatastore_Task, 'duration_secs': 0.022579} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.297938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.297938] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.297938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.297938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.297938] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.298501] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc24f03c-fb43-4680-938e-8270824ba85f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.311242] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.312031] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 980.313029] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52f8cd45-e5e3-4160-8627-5a93aa3cc7a1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.320516] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 980.320516] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523fc0a5-cf9e-e6bb-37f0-9a875edbd677" [ 980.320516] env[61986]: _type = "Task" [ 980.320516] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.331023] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523fc0a5-cf9e-e6bb-37f0-9a875edbd677, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.376483] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.376861] env[61986]: DEBUG nova.objects.instance [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'migration_context' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.492184] env[61986]: DEBUG nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.532725] env[61986]: DEBUG nova.scheduler.client.report [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 109 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 980.533095] env[61986]: DEBUG nova.compute.provider_tree [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 109 to 110 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 980.533172] env[61986]: DEBUG nova.compute.provider_tree [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.721219] env[61986]: DEBUG oslo_vmware.api [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160167, 'name': PowerOffVM_Task, 'duration_secs': 0.23179} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.722143] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 980.722143] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 980.722348] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2186e49b-bb48-46e9-8669-7a293617430d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.772865] env[61986]: DEBUG nova.network.neutron [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Updated VIF entry in instance network info cache for port dd370310-3b48-4932-8a9f-12a6e4298161. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 980.773275] env[61986]: DEBUG nova.network.neutron [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Updating instance_info_cache with network_info: [{"id": "dd370310-3b48-4932-8a9f-12a6e4298161", "address": "fa:16:3e:18:01:e3", "network": {"id": "0f7e1d98-04fa-48fc-957b-ad07ea546de9", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-404545563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ea5fc77300041f5aeccb8773300ef0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd370310-3b", "ovs_interfaceid": "dd370310-3b48-4932-8a9f-12a6e4298161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.789665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.796180] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 980.796432] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 980.796621] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleting the datastore file [datastore1] 3b925174-cf0b-4c6c-b6b6-1c99a50026be {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.796893] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af36b669-2fb5-4811-ba05-f34576c62a20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.805611] env[61986]: DEBUG oslo_vmware.api [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for the task: (returnval){ [ 980.805611] env[61986]: value = "task-1160169" [ 980.805611] env[61986]: _type = "Task" [ 980.805611] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.817869] env[61986]: DEBUG oslo_vmware.api [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.835810] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523fc0a5-cf9e-e6bb-37f0-9a875edbd677, 'name': SearchDatastore_Task, 'duration_secs': 0.016901} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.836684] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d05012f9-bcb7-458a-97c5-0998974f2e2d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.843712] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 980.843712] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a46378-2d3d-1c7b-e2cf-0ad30c6e2736" [ 980.843712] env[61986]: _type = "Task" [ 980.843712] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.856882] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a46378-2d3d-1c7b-e2cf-0ad30c6e2736, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.880707] env[61986]: DEBUG nova.objects.base [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Object Instance<92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f> lazy-loaded attributes: info_cache,migration_context {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 980.882070] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb95ea1-13d5-4a75-a69f-76388a7235b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.914135] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75839e83-383b-42aa-9d14-94452dc422c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.922115] env[61986]: DEBUG oslo_vmware.api [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 980.922115] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c6720f-998d-c82c-9e8f-6f8bd681287f" [ 980.922115] env[61986]: _type = "Task" [ 980.922115] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.934425] env[61986]: DEBUG oslo_vmware.api [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c6720f-998d-c82c-9e8f-6f8bd681287f, 'name': SearchDatastore_Task, 'duration_secs': 0.009686} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.934771] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.000476] env[61986]: DEBUG nova.network.neutron [-] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.039085] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.039828] env[61986]: DEBUG nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.043370] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.621s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.043692] env[61986]: DEBUG nova.objects.instance [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lazy-loading 'resources' on Instance uuid 3d5f4513-bbc4-404e-9d3e-340bd369fc3c {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.057031] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Successfully updated port: 0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.276350] env[61986]: DEBUG oslo_concurrency.lockutils [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] Releasing lock "refresh_cache-73b2c463-0460-4a68-b788-ab0c7818d8f8" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.276677] env[61986]: DEBUG nova.compute.manager [req-5b633716-125b-4f3f-82e5-7eb36b693657 req-259d4a7d-ceec-4c51-ba00-e932219f3bdf service nova] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Received event network-vif-deleted-a9804290-f49e-46f6-90c6-3ce3e529d094 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.322766] env[61986]: DEBUG oslo_vmware.api [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Task: {'id': task-1160169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230432} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.323464] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.323464] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 981.323464] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 981.323675] env[61986]: INFO nova.compute.manager [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Took 1.14 seconds to destroy the instance on the hypervisor. [ 981.323873] env[61986]: DEBUG oslo.service.loopingcall [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.324101] env[61986]: DEBUG nova.compute.manager [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.324755] env[61986]: DEBUG nova.network.neutron [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 981.360955] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a46378-2d3d-1c7b-e2cf-0ad30c6e2736, 'name': SearchDatastore_Task, 'duration_secs': 0.01481} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.361260] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.361518] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 73b2c463-0460-4a68-b788-ab0c7818d8f8/73b2c463-0460-4a68-b788-ab0c7818d8f8.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 981.361792] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d10bd61a-1c7f-4291-a5f5-c69843e7b18f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.372240] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 981.372240] env[61986]: value = "task-1160170" [ 981.372240] env[61986]: _type = "Task" [ 981.372240] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.391964] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160170, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.503190] env[61986]: INFO nova.compute.manager [-] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Took 1.51 seconds to deallocate network for instance. [ 981.547246] env[61986]: DEBUG nova.compute.utils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.552230] env[61986]: DEBUG nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 981.561846] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.562013] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.562181] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.886402] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160170, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.895378] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d54107e-a235-48a3-b80e-175547faf3a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.905252] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6adeb92-0eb6-4cf0-832d-56321c53093d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.942109] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239c81ea-c9fa-4547-9986-fc761d36c66f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.952821] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751afaa1-77d9-40f8-92a8-30104f64899f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.969683] env[61986]: DEBUG nova.compute.provider_tree [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.054703] env[61986]: DEBUG nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.078550] env[61986]: INFO nova.compute.manager [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Took 0.57 seconds to detach 1 volumes for instance. [ 982.081324] env[61986]: DEBUG nova.compute.manager [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Deleting volume: 5ed76af9-6e13-4e9f-aa4d-ecfd170f14f5 {{(pid=61986) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 982.108592] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 982.155076] env[61986]: DEBUG nova.network.neutron [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.386037] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160170, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631121} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.386432] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 73b2c463-0460-4a68-b788-ab0c7818d8f8/73b2c463-0460-4a68-b788-ab0c7818d8f8.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 982.386629] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.387019] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2833e2bf-18eb-48cc-9526-4c207a17ccd1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.398672] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 982.398672] env[61986]: value = "task-1160172" [ 982.398672] env[61986]: _type = "Task" [ 982.398672] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.410795] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160172, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.474402] env[61986]: DEBUG nova.scheduler.client.report [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.633593] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.660749] env[61986]: INFO nova.compute.manager [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Took 1.34 seconds to deallocate network for instance. [ 982.784282] env[61986]: DEBUG nova.network.neutron [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updating instance_info_cache with network_info: [{"id": "759ea478-47fb-45dc-a788-f500f86de7fd", "address": "fa:16:3e:4b:a2:cd", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap759ea478-47", "ovs_interfaceid": "759ea478-47fb-45dc-a788-f500f86de7fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3205584-42e2-4cb4-8b2d-0dd929934176", "address": "fa:16:3e:70:37:2b", "network": {"id": "b7b73ff9-c762-414e-a94f-72865f1c2e1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1787343703", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3205584-42", "ovs_interfaceid": "b3205584-42e2-4cb4-8b2d-0dd929934176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0684bb22-0a77-4633-b834-a66fe18747c5", "address": "fa:16:3e:91:56:27", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0684bb22-0a", "ovs_interfaceid": "0684bb22-0a77-4633-b834-a66fe18747c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.909330] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160172, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.171155} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.909724] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.910669] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea21af99-b0be-4317-8fd4-55d7c7199783 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.934287] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 73b2c463-0460-4a68-b788-ab0c7818d8f8/73b2c463-0460-4a68-b788-ab0c7818d8f8.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.935024] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc4dffd8-f581-435f-9d81-bb31ff15627d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.956891] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 982.956891] env[61986]: value = "task-1160173" [ 982.956891] env[61986]: _type = "Task" [ 982.956891] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.965951] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160173, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.981319] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.938s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.984466] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.264s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.984553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.984712] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 982.985063] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.207s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.985303] env[61986]: DEBUG nova.objects.instance [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lazy-loading 'resources' on Instance uuid 021a27ec-cee0-454d-8daf-e6a82bd9330c {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.987032] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d31576-6276-4ca4-9549-0924c85d35d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.996726] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d38b83-d603-492c-9b5c-2108d9cac832 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.014097] env[61986]: INFO nova.scheduler.client.report [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted allocations for instance 3d5f4513-bbc4-404e-9d3e-340bd369fc3c [ 983.015913] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8d8baa-bfe0-4289-96ea-0758f3a38d51 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.026231] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ec2383-179f-43c1-80dc-a672fdf46660 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.062076] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179099MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 983.062250] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.069796] env[61986]: DEBUG nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.168340] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.287246] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Releasing lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.288512] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Instance network_info: |[{"id": "759ea478-47fb-45dc-a788-f500f86de7fd", "address": "fa:16:3e:4b:a2:cd", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap759ea478-47", "ovs_interfaceid": "759ea478-47fb-45dc-a788-f500f86de7fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3205584-42e2-4cb4-8b2d-0dd929934176", "address": "fa:16:3e:70:37:2b", "network": {"id": "b7b73ff9-c762-414e-a94f-72865f1c2e1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1787343703", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3205584-42", "ovs_interfaceid": "b3205584-42e2-4cb4-8b2d-0dd929934176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0684bb22-0a77-4633-b834-a66fe18747c5", "address": "fa:16:3e:91:56:27", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0684bb22-0a", "ovs_interfaceid": "0684bb22-0a77-4633-b834-a66fe18747c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 983.288783] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:a2:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '030ecc21-dc1c-4283-854e-88e623b3970a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '759ea478-47fb-45dc-a788-f500f86de7fd', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:37:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3205584-42e2-4cb4-8b2d-0dd929934176', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:56:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '030ecc21-dc1c-4283-854e-88e623b3970a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0684bb22-0a77-4633-b834-a66fe18747c5', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.300284] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Creating folder: Project (8c64e7f5b4b64d1988edf2a2fcbe32eb). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 983.301189] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f652a00-a01e-4614-94df-c613ac60072c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.314802] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Created folder: Project (8c64e7f5b4b64d1988edf2a2fcbe32eb) in parent group-v252271. [ 983.314977] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Creating folder: Instances. Parent ref: group-v252425. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 983.315257] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c00290b3-7a59-4528-9e35-011ab240acff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.327641] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Created folder: Instances in parent group-v252425. [ 983.328014] env[61986]: DEBUG oslo.service.loopingcall [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.328254] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 983.328510] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-397996a9-7594-4268-bdff-73b3b2414b49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.356784] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.356784] env[61986]: value = "task-1160176" [ 983.356784] env[61986]: _type = "Task" [ 983.356784] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.366302] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160176, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.469035] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160173, 'name': ReconfigVM_Task, 'duration_secs': 0.350991} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.469035] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 73b2c463-0460-4a68-b788-ab0c7818d8f8/73b2c463-0460-4a68-b788-ab0c7818d8f8.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.469035] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc207364-e50c-40cd-a5ba-22d1b169c2b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.477445] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 983.477445] env[61986]: value = "task-1160177" [ 983.477445] env[61986]: _type = "Task" [ 983.477445] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.487341] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160177, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.524991] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3bd3cbae-de50-41c5-8d22-b4600f97b6ff tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "3d5f4513-bbc4-404e-9d3e-340bd369fc3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.518s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.801565] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bbadae-79b6-4c9c-87e4-07ff50eb2d2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.811762] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ef0933-3337-4630-b7ce-036b84839814 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.849079] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473e0704-86f7-4353-a274-41273fda7190 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.857834] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8484719c-f114-4b12-b99e-4c722d5d1ac7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.875451] env[61986]: DEBUG nova.compute.provider_tree [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.880754] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160176, 'name': CreateVM_Task, 'duration_secs': 0.514095} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.881107] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 983.882275] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.882456] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.882804] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.883102] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d2e6c75-e7fe-46bd-9e43-80702e143a0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.889492] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 983.889492] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c94ce-c71a-9ecc-bb49-b9c1995868f4" [ 983.889492] env[61986]: _type = "Task" [ 983.889492] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.898681] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c94ce-c71a-9ecc-bb49-b9c1995868f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.989877] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160177, 'name': Rename_Task, 'duration_secs': 0.185621} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.990335] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 983.990731] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-771187b4-1d9d-4991-8c66-4c71285da125 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.999909] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 983.999909] env[61986]: value = "task-1160178" [ 983.999909] env[61986]: _type = "Task" [ 983.999909] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.008845] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.380187] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.380498] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.380628] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.380816] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.381074] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.381244] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.381704] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.381704] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.381912] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.381985] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.382166] env[61986]: DEBUG nova.virt.hardware [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.384461] env[61986]: DEBUG nova.scheduler.client.report [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.388570] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fd1672-fa30-4535-8bc1-9a90c9ae7ca9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.406356] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.407021] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.407021] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.407021] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.407217] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.408073] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.408073] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.408073] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.408073] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.408282] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.408324] env[61986]: DEBUG nova.virt.hardware [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.409752] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049e43ef-3778-41a5-933a-22ff11e79c94 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.415050] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c94ce-c71a-9ecc-bb49-b9c1995868f4, 'name': SearchDatastore_Task, 'duration_secs': 0.0166} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.418122] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b44964-8f79-49dc-a506-a54661d251c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.420560] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.420946] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.421098] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.421258] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.421505] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.422774] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b78e7a92-eb2b-40d1-aee0-0dda09297501 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.437668] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254cc62-8e04-d799-2237-cc0523a3ac31/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 984.438755] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.444781] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating folder: Project (6efbe9d217c14411a706c1e13f6bee14). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 984.445882] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2303438d-1153-4b76-a9a3-36480202410f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.452122] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0c0902f-b238-40b2-8bfa-87892e61069b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.456649] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd84d9f-515d-4226-8495-acc32c388b7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.464065] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254cc62-8e04-d799-2237-cc0523a3ac31/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 984.464276] env[61986]: ERROR oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254cc62-8e04-d799-2237-cc0523a3ac31/disk-0.vmdk due to incomplete transfer. [ 984.464534] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.464699] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 984.465776] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c6e2dc53-bd69-40a2-8110-4165ad48eff1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.467443] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92b36d3b-2f87-4907-a2ee-f89cb8d56e4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.479375] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.485119] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating folder: Project (6efbe9d217c14411a706c1e13f6bee14). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 984.487747] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66309e39-0660-44d1-af24-a9682928f003 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.489587] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Created folder: Project (6efbe9d217c14411a706c1e13f6bee14) in parent group-v252271. [ 984.489831] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating folder: Instances. Parent ref: group-v252428. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 984.490573] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ea81b84-3571-458b-b089-f40d2e3d42a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.493741] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 984.493741] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3e8ca-0c7d-24bc-44e6-ae0c05525752" [ 984.493741] env[61986]: _type = "Task" [ 984.493741] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.498390] env[61986]: DEBUG oslo_vmware.rw_handles [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254cc62-8e04-d799-2237-cc0523a3ac31/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 984.498613] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Uploaded image 7d7a5833-c7c3-4049-9eab-4941b2fe99d6 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 984.500334] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 984.501113] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-de6a6daa-9722-4b55-b51b-0ccbed3c7328 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.507028] env[61986]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 984.507221] env[61986]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61986) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 984.512022] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Folder already exists: Project (6efbe9d217c14411a706c1e13f6bee14). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 984.512324] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating folder: Instances. Parent ref: group-v252428. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 984.512635] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3e8ca-0c7d-24bc-44e6-ae0c05525752, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.512917] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Created folder: Instances in parent group-v252428. [ 984.513148] env[61986]: DEBUG oslo.service.loopingcall [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.513712] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4dd252e5-d418-468f-bf62-80faea94559b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.515490] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 984.519663] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-773683fb-50b1-4cd1-9dd6-20c3ba952e17 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.532707] env[61986]: DEBUG oslo_vmware.api [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160178, 'name': PowerOnVM_Task, 'duration_secs': 0.498828} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.533031] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 984.533031] env[61986]: value = "task-1160182" [ 984.533031] env[61986]: _type = "Task" [ 984.533031] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.534701] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 984.535066] env[61986]: INFO nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Took 8.15 seconds to spawn the instance on the hypervisor. [ 984.535372] env[61986]: DEBUG nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.536774] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d059d7-1aac-4500-9203-d427d1ceec9a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.543743] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.543743] env[61986]: value = "task-1160184" [ 984.543743] env[61986]: _type = "Task" [ 984.543743] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.563303] env[61986]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 984.563574] env[61986]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61986) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 984.563884] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Folder already exists: Instances. Parent ref: group-v252428. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 984.564193] env[61986]: DEBUG oslo.service.loopingcall [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.571496] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 984.571715] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160184, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.571958] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160182, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.574912] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3e224c0-0aec-4a1e-beca-2629a4e8aa6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.596146] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.596146] env[61986]: value = "task-1160185" [ 984.596146] env[61986]: _type = "Task" [ 984.596146] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.605347] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160185, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.615529] env[61986]: DEBUG nova.compute.manager [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-vif-plugged-b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.615529] env[61986]: DEBUG oslo_concurrency.lockutils [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] Acquiring lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.615529] env[61986]: DEBUG oslo_concurrency.lockutils [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.615529] env[61986]: DEBUG oslo_concurrency.lockutils [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.615529] env[61986]: DEBUG nova.compute.manager [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] No waiting events found dispatching network-vif-plugged-b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.616125] env[61986]: WARNING nova.compute.manager [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received unexpected event network-vif-plugged-b3205584-42e2-4cb4-8b2d-0dd929934176 for instance with vm_state building and task_state spawning. [ 984.616125] env[61986]: DEBUG nova.compute.manager [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-changed-b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 984.616125] env[61986]: DEBUG nova.compute.manager [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Refreshing instance network info cache due to event network-changed-b3205584-42e2-4cb4-8b2d-0dd929934176. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 984.616125] env[61986]: DEBUG oslo_concurrency.lockutils [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] Acquiring lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.616278] env[61986]: DEBUG oslo_concurrency.lockutils [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] Acquired lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.616448] env[61986]: DEBUG nova.network.neutron [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Refreshing network info cache for port b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.707080] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523df22d-8f8f-838b-1b63-004eb9ab19ea/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 984.708267] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4626ebcd-29ac-4d03-babf-b49cb6aac522 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.715241] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523df22d-8f8f-838b-1b63-004eb9ab19ea/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 984.715470] env[61986]: ERROR oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523df22d-8f8f-838b-1b63-004eb9ab19ea/disk-0.vmdk due to incomplete transfer. [ 984.715788] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f3e60455-bd86-4026-ab2f-9dad89949b44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.724104] env[61986]: DEBUG oslo_vmware.rw_handles [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523df22d-8f8f-838b-1b63-004eb9ab19ea/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 984.724326] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Uploaded image 24ded09f-c9f0-435c-93fc-1aef721bbca3 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 984.726777] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 984.727430] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-96855129-2b4c-4847-b0d3-c3b1ade39174 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.734804] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 984.734804] env[61986]: value = "task-1160186" [ 984.734804] env[61986]: _type = "Task" [ 984.734804] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.746471] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160186, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.894562] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.895887] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.968s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.897640] env[61986]: INFO nova.compute.claims [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.925039] env[61986]: INFO nova.scheduler.client.report [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Deleted allocations for instance 021a27ec-cee0-454d-8daf-e6a82bd9330c [ 984.966058] env[61986]: DEBUG nova.compute.manager [req-adf98441-1760-4f79-bac6-9989567c3658 req-21f90eb9-91f0-497f-81ee-47693fbd67c5 service nova] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Received event network-vif-deleted-8157a838-3c36-4936-a187-460e14d7f129 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.007185] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b3e8ca-0c7d-24bc-44e6-ae0c05525752, 'name': SearchDatastore_Task, 'duration_secs': 0.025387} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.008139] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e27037e5-683f-4ab0-a0f3-9a849750323d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.015473] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 985.015473] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5218b34e-33cb-e0ab-2f8c-e2d639d12e0f" [ 985.015473] env[61986]: _type = "Task" [ 985.015473] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.024175] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5218b34e-33cb-e0ab-2f8c-e2d639d12e0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.047930] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160182, 'name': Destroy_Task, 'duration_secs': 0.436139} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.051200] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Destroyed the VM [ 985.051430] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 985.052239] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-57064608-fc10-4f93-9103-b1812c2abe13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.059472] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160184, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.061149] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 985.061149] env[61986]: value = "task-1160187" [ 985.061149] env[61986]: _type = "Task" [ 985.061149] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.071371] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160187, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.098577] env[61986]: INFO nova.compute.manager [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Took 26.72 seconds to build instance. [ 985.106843] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160185, 'name': CreateVM_Task, 'duration_secs': 0.387623} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.106843] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 985.107422] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.107769] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.108629] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.108629] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dddcc122-6fcc-49f1-abea-180f4143c229 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.115449] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 985.115449] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5233af35-64b9-7a9b-6fcf-7c19e49aee54" [ 985.115449] env[61986]: _type = "Task" [ 985.115449] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.127795] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5233af35-64b9-7a9b-6fcf-7c19e49aee54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.247274] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160186, 'name': Destroy_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.432264] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4fad0f93-a350-4207-906e-9d55b6d662e3 tempest-ServerRescueTestJSON-2105203801 tempest-ServerRescueTestJSON-2105203801-project-member] Lock "021a27ec-cee0-454d-8daf-e6a82bd9330c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.764s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.527390] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5218b34e-33cb-e0ab-2f8c-e2d639d12e0f, 'name': SearchDatastore_Task, 'duration_secs': 0.034857} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.527870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.528054] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 92077300-2e56-4dec-9e8d-9b566ac3fe0e/92077300-2e56-4dec-9e8d-9b566ac3fe0e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 985.528341] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2318b099-928c-4747-84e9-0706e7bb5df6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.533520] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.539607] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 985.539607] env[61986]: value = "task-1160188" [ 985.539607] env[61986]: _type = "Task" [ 985.539607] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.565766] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160184, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.571888] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160188, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.577986] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160187, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.601074] env[61986]: DEBUG oslo_concurrency.lockutils [None req-01b17723-42cb-4b41-ab9e-01ea3054ebb3 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.233s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.607830] env[61986]: DEBUG nova.network.neutron [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updated VIF entry in instance network info cache for port b3205584-42e2-4cb4-8b2d-0dd929934176. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.608369] env[61986]: DEBUG nova.network.neutron [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updating instance_info_cache with network_info: [{"id": "759ea478-47fb-45dc-a788-f500f86de7fd", "address": "fa:16:3e:4b:a2:cd", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap759ea478-47", "ovs_interfaceid": "759ea478-47fb-45dc-a788-f500f86de7fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3205584-42e2-4cb4-8b2d-0dd929934176", "address": "fa:16:3e:70:37:2b", "network": {"id": "b7b73ff9-c762-414e-a94f-72865f1c2e1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1787343703", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3205584-42", "ovs_interfaceid": "b3205584-42e2-4cb4-8b2d-0dd929934176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0684bb22-0a77-4633-b834-a66fe18747c5", "address": "fa:16:3e:91:56:27", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0684bb22-0a", "ovs_interfaceid": "0684bb22-0a77-4633-b834-a66fe18747c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.627146] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5233af35-64b9-7a9b-6fcf-7c19e49aee54, 'name': SearchDatastore_Task, 'duration_secs': 0.01975} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.627451] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.627678] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.627929] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.628105] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.628292] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.628553] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d886b56-92d8-4cf1-ac1b-284979b249c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.639322] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.639553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.639763] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.639972] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.640180] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.642801] env[61986]: INFO nova.compute.manager [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Terminating instance [ 985.644647] env[61986]: DEBUG nova.compute.manager [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.644842] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 985.645680] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bdcdef-27dd-4f82-980f-4e670c7c907d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.650101] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.650279] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 985.651018] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-249e101c-673c-4643-acc6-3227b78b73a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.656983] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.657590] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f35df620-d2c7-42f9-acd7-441fd237660b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.660836] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 985.660836] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f320af-a91f-fe85-ade0-450d3a98020d" [ 985.660836] env[61986]: _type = "Task" [ 985.660836] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.666101] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 985.666101] env[61986]: value = "task-1160189" [ 985.666101] env[61986]: _type = "Task" [ 985.666101] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.669007] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f320af-a91f-fe85-ade0-450d3a98020d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.676410] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.746095] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160186, 'name': Destroy_Task, 'duration_secs': 0.850012} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.746413] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Destroyed the VM [ 985.746609] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 985.746866] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9b57fff2-5240-4aa9-8825-d5da54f87f47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.753927] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 985.753927] env[61986]: value = "task-1160190" [ 985.753927] env[61986]: _type = "Task" [ 985.753927] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.763023] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160190, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.050395] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160188, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.066625] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160184, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.080108] env[61986]: DEBUG oslo_vmware.api [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160187, 'name': RemoveSnapshot_Task, 'duration_secs': 0.531177} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.080443] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 986.080720] env[61986]: INFO nova.compute.manager [None req-a04611d4-ee8c-4dc1-b6ee-98cfe85a778b tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Took 13.92 seconds to snapshot the instance on the hypervisor. [ 986.111351] env[61986]: DEBUG oslo_concurrency.lockutils [req-fb2b632b-e696-4254-bdd7-8aadc96ff835 req-9c9c3863-3263-417b-a975-e302d92ba43e service nova] Releasing lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.177498] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f320af-a91f-fe85-ade0-450d3a98020d, 'name': SearchDatastore_Task, 'duration_secs': 0.02392} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.182177] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72bcc1c1-2e2b-45f9-988a-a5fcc16b9fd7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.188406] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.193737] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 986.193737] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f2ee5d-c3cf-e1a9-c901-c2411841cfd8" [ 986.193737] env[61986]: _type = "Task" [ 986.193737] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.210159] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f2ee5d-c3cf-e1a9-c901-c2411841cfd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.272534] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160190, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.308514] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb7e0b7-2c7a-459a-aaf5-577ae8164ea2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.320602] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de70646-4151-4dfd-98d0-06b3c2c2c49d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.366904] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f80587-0750-42b0-a6d3-c9c2a5826900 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.379260] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b434505-3088-4570-8fff-93a48f225dab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.398804] env[61986]: DEBUG nova.compute.provider_tree [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.520839] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "73b2c463-0460-4a68-b788-ab0c7818d8f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.521229] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.521454] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "73b2c463-0460-4a68-b788-ab0c7818d8f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.521649] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.521824] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.524095] env[61986]: INFO nova.compute.manager [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Terminating instance [ 986.526053] env[61986]: DEBUG nova.compute.manager [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.526262] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 986.527105] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c434b1-d18d-46c5-9073-167a2f8c99a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.538092] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 986.538383] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ef2e7a6-e6fc-407d-a475-0f7835b83a2c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.558722] env[61986]: DEBUG oslo_vmware.api [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 986.558722] env[61986]: value = "task-1160191" [ 986.558722] env[61986]: _type = "Task" [ 986.558722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.566236] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160188, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.882688} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.566595] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160184, 'name': CreateVM_Task, 'duration_secs': 1.635385} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.570000] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 92077300-2e56-4dec-9e8d-9b566ac3fe0e/92077300-2e56-4dec-9e8d-9b566ac3fe0e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 986.570327] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.570561] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 986.570894] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1cad89b-95f9-4ddd-a050-3cbd64c040c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.573350] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.573484] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.573742] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.574588] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb7400db-3c9d-4891-ad75-66bb44310de9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.580227] env[61986]: DEBUG oslo_vmware.api [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160191, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.586236] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 986.586236] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52923a4e-8622-d8e0-85e2-08361bfd8237" [ 986.586236] env[61986]: _type = "Task" [ 986.586236] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.589562] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 986.589562] env[61986]: value = "task-1160192" [ 986.589562] env[61986]: _type = "Task" [ 986.589562] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.600894] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52923a4e-8622-d8e0-85e2-08361bfd8237, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.606928] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160192, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.681480] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160189, 'name': PowerOffVM_Task, 'duration_secs': 0.727904} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.681791] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 986.682013] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 986.682354] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2c5602e-a669-47b4-92ec-b70c9db0db4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.703612] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-vif-plugged-0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.704016] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquiring lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.704272] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.704502] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.704697] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] No waiting events found dispatching network-vif-plugged-0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.704870] env[61986]: WARNING nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received unexpected event network-vif-plugged-0684bb22-0a77-4633-b834-a66fe18747c5 for instance with vm_state building and task_state spawning. [ 986.705095] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-changed-0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.705286] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Refreshing instance network info cache due to event network-changed-0684bb22-0a77-4633-b834-a66fe18747c5. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 986.705480] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquiring lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.705702] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquired lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.705889] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Refreshing network info cache for port 0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.714309] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f2ee5d-c3cf-e1a9-c901-c2411841cfd8, 'name': SearchDatastore_Task, 'duration_secs': 0.061989} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.714652] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.714917] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 986.715218] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3dc38a0-fd62-4239-a422-31d1a2b21e03 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.726589] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 986.726589] env[61986]: value = "task-1160194" [ 986.726589] env[61986]: _type = "Task" [ 986.726589] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.737194] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160194, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.763520] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 986.763881] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 986.764207] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleting the datastore file [datastore2] b9706e51-0f74-4dbc-aab1-ea640ac78c62 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.768674] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f84e7e9f-0d6a-4601-a2fa-ff67fe5d4572 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.770828] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160190, 'name': RemoveSnapshot_Task, 'duration_secs': 0.978888} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.771212] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 986.771594] env[61986]: DEBUG nova.compute.manager [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.774158] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97225df7-68eb-4deb-951f-bad99b23bf9b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.777311] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 986.777311] env[61986]: value = "task-1160195" [ 986.777311] env[61986]: _type = "Task" [ 986.777311] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.791036] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160195, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.903022] env[61986]: DEBUG nova.scheduler.client.report [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.080180] env[61986]: DEBUG oslo_vmware.api [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160191, 'name': PowerOffVM_Task, 'duration_secs': 0.237156} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.080531] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.080734] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 987.081061] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dc185a2-39b8-4a29-9b8b-70d62fe72aaf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.097304] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52923a4e-8622-d8e0-85e2-08361bfd8237, 'name': SearchDatastore_Task, 'duration_secs': 0.019818} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.101044] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.101394] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.101653] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.101806] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.101993] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.102580] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49341f6e-1a19-4b0b-bb23-bfe1f1218689 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.113346] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160192, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083022} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.113634] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.114523] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c042516-afa8-477a-98dc-8f5315a63cfe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.120781] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.121197] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 987.137149] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f18262e-9d87-44e1-a1da-013cdbb3b9de {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.149086] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 92077300-2e56-4dec-9e8d-9b566ac3fe0e/92077300-2e56-4dec-9e8d-9b566ac3fe0e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.149490] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e359f5f-27ef-4b78-92c5-47fb6fe7f183 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.168366] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 987.168366] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 987.168366] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleting the datastore file [datastore1] 73b2c463-0460-4a68-b788-ab0c7818d8f8 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.172131] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a51b28f-e382-44bf-8b53-09f0f6646f41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.173667] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 987.173667] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d0f9cf-797c-d8df-335c-2ec3f453fe5a" [ 987.173667] env[61986]: _type = "Task" [ 987.173667] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.175347] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 987.175347] env[61986]: value = "task-1160197" [ 987.175347] env[61986]: _type = "Task" [ 987.175347] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.185025] env[61986]: DEBUG oslo_vmware.api [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for the task: (returnval){ [ 987.185025] env[61986]: value = "task-1160198" [ 987.185025] env[61986]: _type = "Task" [ 987.185025] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.197237] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160197, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.197535] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d0f9cf-797c-d8df-335c-2ec3f453fe5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.205106] env[61986]: DEBUG oslo_vmware.api [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160198, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.240957] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160194, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.291644] env[61986]: DEBUG oslo_vmware.api [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160195, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244839} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.291949] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.292201] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 987.292434] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 987.292615] env[61986]: INFO nova.compute.manager [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Took 1.65 seconds to destroy the instance on the hypervisor. [ 987.292867] env[61986]: DEBUG oslo.service.loopingcall [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.293087] env[61986]: DEBUG nova.compute.manager [-] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.293150] env[61986]: DEBUG nova.network.neutron [-] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 987.295568] env[61986]: INFO nova.compute.manager [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Shelve offloading [ 987.299220] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 987.299473] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e857e7e3-8a54-4d15-9412-ace22fdec563 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.310997] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 987.310997] env[61986]: value = "task-1160199" [ 987.310997] env[61986]: _type = "Task" [ 987.310997] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.324938] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 987.325212] env[61986]: DEBUG nova.compute.manager [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.326109] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1621ad87-200d-4da5-9d99-d360e9ffc048 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.333709] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.333887] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.334067] env[61986]: DEBUG nova.network.neutron [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.408876] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.409589] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.412904] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.168s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.413123] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.417015] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.626s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.417015] env[61986]: DEBUG nova.objects.instance [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'resources' on Instance uuid f15bf19d-b86a-4b0a-ac1a-9df8e77e9382 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.462674] env[61986]: INFO nova.scheduler.client.report [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocations for instance f42e92ef-8a90-473c-8662-57cbb9e3f4ee [ 987.606405] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updated VIF entry in instance network info cache for port 0684bb22-0a77-4633-b834-a66fe18747c5. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.607136] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updating instance_info_cache with network_info: [{"id": "759ea478-47fb-45dc-a788-f500f86de7fd", "address": "fa:16:3e:4b:a2:cd", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap759ea478-47", "ovs_interfaceid": "759ea478-47fb-45dc-a788-f500f86de7fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b3205584-42e2-4cb4-8b2d-0dd929934176", "address": "fa:16:3e:70:37:2b", "network": {"id": "b7b73ff9-c762-414e-a94f-72865f1c2e1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1787343703", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.237", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3205584-42", "ovs_interfaceid": "b3205584-42e2-4cb4-8b2d-0dd929934176", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0684bb22-0a77-4633-b834-a66fe18747c5", "address": "fa:16:3e:91:56:27", "network": {"id": "885bf1b0-ee4b-49ec-b6f0-d7ff6711a4b4", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2025508067", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0684bb22-0a", "ovs_interfaceid": "0684bb22-0a77-4633-b834-a66fe18747c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.700094] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d0f9cf-797c-d8df-335c-2ec3f453fe5a, 'name': SearchDatastore_Task, 'duration_secs': 0.061545} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.707862] env[61986]: DEBUG nova.compute.manager [req-164d5b10-6445-472a-8a3c-a09608271c28 req-0766c869-5615-4bb5-9b90-c9672b1c9fdc service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Received event network-vif-deleted-1234061c-4805-4e9a-9993-3e09549f5569 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 987.708096] env[61986]: INFO nova.compute.manager [req-164d5b10-6445-472a-8a3c-a09608271c28 req-0766c869-5615-4bb5-9b90-c9672b1c9fdc service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Neutron deleted interface 1234061c-4805-4e9a-9993-3e09549f5569; detaching it from the instance and deleting it from the info cache [ 987.708441] env[61986]: DEBUG nova.network.neutron [req-164d5b10-6445-472a-8a3c-a09608271c28 req-0766c869-5615-4bb5-9b90-c9672b1c9fdc service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.710767] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160197, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.714151] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3a2ca63-ca64-4ece-80a9-a2f13be959ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.717228] env[61986]: DEBUG oslo_vmware.api [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Task: {'id': task-1160198, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.297904} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.717484] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.717667] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 987.717850] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 987.718108] env[61986]: INFO nova.compute.manager [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Took 1.19 seconds to destroy the instance on the hypervisor. [ 987.718300] env[61986]: DEBUG oslo.service.loopingcall [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.718970] env[61986]: DEBUG nova.compute.manager [-] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.719105] env[61986]: DEBUG nova.network.neutron [-] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 987.723112] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 987.723112] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52db3e53-a25b-c9d2-067f-5843c5434c6d" [ 987.723112] env[61986]: _type = "Task" [ 987.723112] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.735389] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52db3e53-a25b-c9d2-067f-5843c5434c6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.741558] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160194, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.583348} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.741831] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 987.742898] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.742898] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7854425f-f60a-4fab-80f7-b8c7c8948f5f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.750738] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 987.750738] env[61986]: value = "task-1160200" [ 987.750738] env[61986]: _type = "Task" [ 987.750738] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.761597] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.921033] env[61986]: DEBUG nova.compute.utils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.924965] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.927037] env[61986]: DEBUG nova.network.neutron [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 987.977486] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2956d190-bc03-47e8-8e6c-dce774307afb tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "f42e92ef-8a90-473c-8662-57cbb9e3f4ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.643s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.989836] env[61986]: DEBUG nova.policy [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ea19b4c355d4b2797683d78e82eadea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0110dfcd30de444badf6fe685042ba3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 988.110332] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Releasing lock "refresh_cache-92077300-2e56-4dec-9e8d-9b566ac3fe0e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.110617] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.110786] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing instance network info cache due to event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 988.111053] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.111214] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.111381] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.167117] env[61986]: DEBUG nova.network.neutron [-] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.179415] env[61986]: DEBUG nova.network.neutron [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.193813] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160197, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.211366] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4ee1802-fd80-4bdd-bc92-1ea57ff6c960 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.221483] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff47ae6-9a33-42c5-8339-26234461a4dc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.248846] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52db3e53-a25b-c9d2-067f-5843c5434c6d, 'name': SearchDatastore_Task, 'duration_secs': 0.037997} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.251521] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.251987] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 45d80bab-0140-4473-9dc2-4a5481838aed/45d80bab-0140-4473-9dc2-4a5481838aed.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 988.252418] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ce8df63-a025-47fb-b283-5a73d8b879f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.269783] env[61986]: DEBUG nova.compute.manager [req-164d5b10-6445-472a-8a3c-a09608271c28 req-0766c869-5615-4bb5-9b90-c9672b1c9fdc service nova] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Detach interface failed, port_id=1234061c-4805-4e9a-9993-3e09549f5569, reason: Instance b9706e51-0f74-4dbc-aab1-ea640ac78c62 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 988.279874] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198396} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.281376] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.281899] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 988.281899] env[61986]: value = "task-1160201" [ 988.281899] env[61986]: _type = "Task" [ 988.281899] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.282702] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2781d7-b196-4798-a9fc-74b1e00b7969 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.290540] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04246a56-f0c7-4111-9057-795c910472c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.314680] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.320183] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cd85949-3331-459e-8e8b-018384ea1b8e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.335573] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160201, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.336515] env[61986]: DEBUG nova.network.neutron [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Successfully created port: 9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 988.339131] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-bc93bdd5-256e-4900-a9ae-609b0a7d402d-7228964a-4136-4e76-8fe4-a21ab14deacf" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.339131] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bc93bdd5-256e-4900-a9ae-609b0a7d402d-7228964a-4136-4e76-8fe4-a21ab14deacf" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.339131] env[61986]: DEBUG nova.objects.instance [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'flavor' on Instance uuid bc93bdd5-256e-4900-a9ae-609b0a7d402d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.343320] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03228c6b-266f-4bdd-9a47-f353a1f022d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.380750] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f06c69d-f17b-4467-88cd-17140f4fe94a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.384358] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 988.384358] env[61986]: value = "task-1160202" [ 988.384358] env[61986]: _type = "Task" [ 988.384358] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.393805] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee69da6f-4d96-44d9-a519-ab907b3b09d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.403840] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.418673] env[61986]: DEBUG nova.compute.provider_tree [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.425413] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 988.587811] env[61986]: DEBUG nova.network.neutron [-] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.671888] env[61986]: INFO nova.compute.manager [-] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Took 1.38 seconds to deallocate network for instance. [ 988.683233] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.700068] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160197, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.795134] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160201, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496244} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.797372] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 45d80bab-0140-4473-9dc2-4a5481838aed/45d80bab-0140-4473-9dc2-4a5481838aed.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 988.797607] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.797922] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-483311a4-96a1-4df4-9253-a6440d989f65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.807634] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 988.807634] env[61986]: value = "task-1160203" [ 988.807634] env[61986]: _type = "Task" [ 988.807634] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.816888] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160203, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.896953] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160202, 'name': ReconfigVM_Task, 'duration_secs': 0.391046} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.897411] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.898052] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d283468d-2147-445d-90cc-d48631689772 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.905135] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 988.905135] env[61986]: value = "task-1160204" [ 988.905135] env[61986]: _type = "Task" [ 988.905135] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.915670] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160204, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.960408] env[61986]: DEBUG nova.scheduler.client.report [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 110 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 988.960797] env[61986]: DEBUG nova.compute.provider_tree [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 110 to 111 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 988.961077] env[61986]: DEBUG nova.compute.provider_tree [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.016183] env[61986]: DEBUG nova.objects.instance [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'pci_requests' on Instance uuid bc93bdd5-256e-4900-a9ae-609b0a7d402d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.060703] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 989.061794] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ca541f-5f73-4b9f-90ad-58a79ff47e6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.069764] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 989.072464] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56b5db8f-b537-4190-a30a-53cf7ce75d2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.091499] env[61986]: INFO nova.compute.manager [-] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Took 1.37 seconds to deallocate network for instance. [ 989.151160] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 989.151397] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 989.151584] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleting the datastore file [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.151861] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c2282f1-0a4e-47a2-81c4-1bbbb1ce9c5a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.161111] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 989.161111] env[61986]: value = "task-1160206" [ 989.161111] env[61986]: _type = "Task" [ 989.161111] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.171219] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160206, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.187490] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.191807] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated VIF entry in instance network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.192284] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.199498] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160197, 'name': ReconfigVM_Task, 'duration_secs': 1.865983} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.199846] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 92077300-2e56-4dec-9e8d-9b566ac3fe0e/92077300-2e56-4dec-9e8d-9b566ac3fe0e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.200583] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13b5337d-e2b3-4c30-96a7-64cd835a326d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.208097] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 989.208097] env[61986]: value = "task-1160207" [ 989.208097] env[61986]: _type = "Task" [ 989.208097] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.217528] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160207, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.319215] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160203, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069846} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.319572] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.320467] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6fab84-8827-4d7c-8cc4-8f655e92e898 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.343524] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 45d80bab-0140-4473-9dc2-4a5481838aed/45d80bab-0140-4473-9dc2-4a5481838aed.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.343857] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fc34af2-1e39-4e55-b626-ef0179154fd0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.366649] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 989.366649] env[61986]: value = "task-1160208" [ 989.366649] env[61986]: _type = "Task" [ 989.366649] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.376959] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.415633] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160204, 'name': Rename_Task, 'duration_secs': 0.156442} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.415935] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.416203] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2e36cad-c35f-4ad6-ba0e-876cfe9c50c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.423538] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 989.423538] env[61986]: value = "task-1160209" [ 989.423538] env[61986]: _type = "Task" [ 989.423538] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.432689] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160209, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.434798] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 989.462604] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.462909] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.463097] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.463315] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.463479] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.463646] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.463888] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.464072] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.464297] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.464459] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.464640] env[61986]: DEBUG nova.virt.hardware [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.465596] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0409a062-4ca9-4686-9b3c-e41bd5633786 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.469249] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.054s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.471755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.537s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.482376] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82748ff8-09a3-4225-896c-8f4a1d9b44c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.498545] env[61986]: INFO nova.scheduler.client.report [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted allocations for instance f15bf19d-b86a-4b0a-ac1a-9df8e77e9382 [ 989.519519] env[61986]: DEBUG nova.objects.base [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 989.519754] env[61986]: DEBUG nova.network.neutron [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.598418] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.633684] env[61986]: DEBUG nova.policy [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 989.673560] env[61986]: DEBUG oslo_vmware.api [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160206, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253617} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.673905] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.674253] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 989.674574] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 989.692330] env[61986]: INFO nova.scheduler.client.report [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted allocations for instance ee47cab9-5dfd-48ce-ba70-cb800d735b19 [ 989.697165] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.697493] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.697840] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing instance network info cache due to event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 989.698120] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.698490] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.698841] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.718117] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160207, 'name': Rename_Task, 'duration_secs': 0.186839} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.718449] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.718737] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbd8e7fe-44c4-4d37-8f8d-0a38defe723f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.725891] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 989.725891] env[61986]: value = "task-1160210" [ 989.725891] env[61986]: _type = "Task" [ 989.725891] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.735013] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.765877] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "6488f68e-7b84-4462-aef5-25d02db504f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.766220] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "6488f68e-7b84-4462-aef5-25d02db504f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.766387] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "6488f68e-7b84-4462-aef5-25d02db504f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.766545] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "6488f68e-7b84-4462-aef5-25d02db504f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.766709] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "6488f68e-7b84-4462-aef5-25d02db504f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.770816] env[61986]: INFO nova.compute.manager [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Terminating instance [ 989.773044] env[61986]: DEBUG nova.compute.manager [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.773261] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 989.774103] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4249b7a5-ec22-4800-84ba-7e4f28121dd9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.782592] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 989.782890] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-793f6834-da9c-4698-b879-32e14acb3f2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.790517] env[61986]: DEBUG oslo_vmware.api [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 989.790517] env[61986]: value = "task-1160211" [ 989.790517] env[61986]: _type = "Task" [ 989.790517] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.798703] env[61986]: DEBUG oslo_vmware.api [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160211, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.878483] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160208, 'name': ReconfigVM_Task, 'duration_secs': 0.324652} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.878989] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 45d80bab-0140-4473-9dc2-4a5481838aed/45d80bab-0140-4473-9dc2-4a5481838aed.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.880082] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c2d37a8d-25bc-449f-ab7c-aba1445b0f38 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.890025] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 989.890025] env[61986]: value = "task-1160212" [ 989.890025] env[61986]: _type = "Task" [ 989.890025] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.899631] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160212, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.933947] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160209, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.955033] env[61986]: DEBUG nova.network.neutron [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Successfully updated port: 9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.006775] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6640688e-039a-4d29-85e8-b3580203209e tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "f15bf19d-b86a-4b0a-ac1a-9df8e77e9382" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.594s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.045838] env[61986]: DEBUG nova.compute.manager [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Received event network-vif-deleted-dd370310-3b48-4932-8a9f-12a6e4298161 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.046133] env[61986]: DEBUG nova.compute.manager [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-vif-unplugged-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.046663] env[61986]: DEBUG oslo_concurrency.lockutils [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.046811] env[61986]: DEBUG oslo_concurrency.lockutils [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.047169] env[61986]: DEBUG oslo_concurrency.lockutils [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.047607] env[61986]: DEBUG nova.compute.manager [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] No waiting events found dispatching network-vif-unplugged-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.047803] env[61986]: WARNING nova.compute.manager [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received unexpected event network-vif-unplugged-78729401-474d-4e4f-8159-aff2e908be0b for instance with vm_state shelved_offloaded and task_state None. [ 990.048138] env[61986]: DEBUG nova.compute.manager [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-changed-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.048446] env[61986]: DEBUG nova.compute.manager [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Refreshing instance network info cache due to event network-changed-78729401-474d-4e4f-8159-aff2e908be0b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 990.048811] env[61986]: DEBUG oslo_concurrency.lockutils [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.049111] env[61986]: DEBUG oslo_concurrency.lockutils [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.049502] env[61986]: DEBUG nova.network.neutron [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Refreshing network info cache for port 78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 990.198092] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.239853] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160210, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.246389] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6456658f-4363-4c98-998e-66e0b7b46a6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.254148] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c483e784-5b42-4d3c-ab67-e3a5b2551431 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.287569] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51d49a9-4e14-4291-8098-c360473a6290 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.300654] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fd3439-2e85-4bb7-b771-8e87ae410501 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.304675] env[61986]: DEBUG oslo_vmware.api [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160211, 'name': PowerOffVM_Task, 'duration_secs': 0.457859} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.304938] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 990.305123] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 990.305722] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9c77f8d-16e6-4928-af1d-58df5b3cb468 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.316617] env[61986]: DEBUG nova.compute.provider_tree [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.380372] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 990.380490] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 990.380709] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Deleting the datastore file [datastore2] 6488f68e-7b84-4462-aef5-25d02db504f6 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.383538] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6a18373-f4a7-4a2f-9200-6046e2ddd5e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.395074] env[61986]: DEBUG oslo_vmware.api [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 990.395074] env[61986]: value = "task-1160214" [ 990.395074] env[61986]: _type = "Task" [ 990.395074] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.402088] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160212, 'name': Rename_Task, 'duration_secs': 0.268937} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.402831] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 990.403131] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3f4b4c5-b8ff-4b62-a896-0d676b303010 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.407921] env[61986]: DEBUG oslo_vmware.api [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.414392] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 990.414392] env[61986]: value = "task-1160215" [ 990.414392] env[61986]: _type = "Task" [ 990.414392] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.427079] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.440799] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160209, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.459776] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.460553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.460553] env[61986]: DEBUG nova.network.neutron [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 990.503987] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updated VIF entry in instance network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.504393] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.737700] env[61986]: DEBUG oslo_vmware.api [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160210, 'name': PowerOnVM_Task, 'duration_secs': 0.70314} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.738023] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.738238] env[61986]: INFO nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Took 16.41 seconds to spawn the instance on the hypervisor. [ 990.738424] env[61986]: DEBUG nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.739429] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5b051f-aca2-4e76-b06c-70d20339b504 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.765353] env[61986]: DEBUG nova.network.neutron [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updated VIF entry in instance network info cache for port 78729401-474d-4e4f-8159-aff2e908be0b. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.765353] env[61986]: DEBUG nova.network.neutron [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap78729401-47", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.820617] env[61986]: DEBUG nova.scheduler.client.report [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.906787] env[61986]: DEBUG oslo_vmware.api [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169793} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.907083] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.907279] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 990.907472] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 990.907653] env[61986]: INFO nova.compute.manager [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 990.907950] env[61986]: DEBUG oslo.service.loopingcall [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.908169] env[61986]: DEBUG nova.compute.manager [-] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.908268] env[61986]: DEBUG nova.network.neutron [-] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 990.924827] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160215, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.935876] env[61986]: DEBUG oslo_vmware.api [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160209, 'name': PowerOnVM_Task, 'duration_secs': 1.149758} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.936311] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.936556] env[61986]: INFO nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Took 7.87 seconds to spawn the instance on the hypervisor. [ 990.936813] env[61986]: DEBUG nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.938038] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e1b345-a4a7-4400-aa4e-bb10cdae6595 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.007227] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.007514] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Received event network-vif-deleted-a6a66909-ff92-4f28-a4c4-c3811691128d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.007709] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.007906] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing instance network info cache due to event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 991.008182] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.008361] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.008511] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.010276] env[61986]: DEBUG nova.network.neutron [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 991.234523] env[61986]: DEBUG nova.network.neutron [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance_info_cache with network_info: [{"id": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "address": "fa:16:3e:c3:f2:46", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f71edfb-58", "ovs_interfaceid": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.258080] env[61986]: INFO nova.compute.manager [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Took 33.82 seconds to build instance. [ 991.270209] env[61986]: DEBUG oslo_concurrency.lockutils [req-2eb7fb7a-3c68-401f-b786-1582aa686b94 req-c390f019-0cbe-48e0-9bce-7711570029e8 service nova] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.403233] env[61986]: DEBUG nova.network.neutron [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Successfully updated port: 7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.427779] env[61986]: DEBUG oslo_vmware.api [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160215, 'name': PowerOnVM_Task, 'duration_secs': 0.99135} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.428059] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 991.428295] env[61986]: INFO nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Took 10.94 seconds to spawn the instance on the hypervisor. [ 991.428508] env[61986]: DEBUG nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.429337] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23c5580-a140-4be2-a58c-7c1f38b8b743 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.457954] env[61986]: DEBUG oslo_concurrency.lockutils [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.458270] env[61986]: DEBUG oslo_concurrency.lockutils [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.462521] env[61986]: INFO nova.compute.manager [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Took 27.22 seconds to build instance. [ 991.732706] env[61986]: DEBUG nova.compute.manager [req-efd50d96-2ae0-41a7-b30e-d46399364be3 req-0795ac0e-ddb1-4425-97e9-f9ca1f981f8a service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-vif-plugged-7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.732937] env[61986]: DEBUG oslo_concurrency.lockutils [req-efd50d96-2ae0-41a7-b30e-d46399364be3 req-0795ac0e-ddb1-4425-97e9-f9ca1f981f8a service nova] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.733163] env[61986]: DEBUG oslo_concurrency.lockutils [req-efd50d96-2ae0-41a7-b30e-d46399364be3 req-0795ac0e-ddb1-4425-97e9-f9ca1f981f8a service nova] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.733337] env[61986]: DEBUG oslo_concurrency.lockutils [req-efd50d96-2ae0-41a7-b30e-d46399364be3 req-0795ac0e-ddb1-4425-97e9-f9ca1f981f8a service nova] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.733541] env[61986]: DEBUG nova.compute.manager [req-efd50d96-2ae0-41a7-b30e-d46399364be3 req-0795ac0e-ddb1-4425-97e9-f9ca1f981f8a service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] No waiting events found dispatching network-vif-plugged-7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.733676] env[61986]: WARNING nova.compute.manager [req-efd50d96-2ae0-41a7-b30e-d46399364be3 req-0795ac0e-ddb1-4425-97e9-f9ca1f981f8a service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received unexpected event network-vif-plugged-7228964a-4136-4e76-8fe4-a21ab14deacf for instance with vm_state active and task_state None. [ 991.737755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.738079] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Instance network_info: |[{"id": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "address": "fa:16:3e:c3:f2:46", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f71edfb-58", "ovs_interfaceid": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.738664] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:f2:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f71edfb-58a5-43cb-9d35-62cda5bf87ad', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.746791] env[61986]: DEBUG oslo.service.loopingcall [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.747217] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 991.747445] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b188f8e1-3c20-408e-8ccc-ae641a14001e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.764199] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a7008a56-f8ad-4bd5-89c8-a113d6a7334e tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.349s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.771333] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.771333] env[61986]: value = "task-1160216" [ 991.771333] env[61986]: _type = "Task" [ 991.771333] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.774504] env[61986]: DEBUG nova.network.neutron [-] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.781424] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160216, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.802469] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updated VIF entry in instance network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 991.802827] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.831596] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.360s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.831844] env[61986]: DEBUG nova.compute.manager [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61986) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 991.836014] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.202s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.836425] env[61986]: DEBUG nova.objects.instance [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lazy-loading 'resources' on Instance uuid 5170aa51-3307-42b1-b0dd-645dd4036e5b {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.907722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.908017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.908274] env[61986]: DEBUG nova.network.neutron [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.947078] env[61986]: INFO nova.compute.manager [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Took 28.60 seconds to build instance. [ 991.963939] env[61986]: INFO nova.compute.manager [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Detaching volume 853747f9-2ca2-43a2-acfb-5d955470ace1 [ 991.966800] env[61986]: DEBUG oslo_concurrency.lockutils [None req-56c441ba-65ee-4f34-950c-16c733304be8 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.736s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.000538] env[61986]: INFO nova.virt.block_device [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Attempting to driver detach volume 853747f9-2ca2-43a2-acfb-5d955470ace1 from mountpoint /dev/sdb [ 992.000781] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 992.001045] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252367', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'name': 'volume-853747f9-2ca2-43a2-acfb-5d955470ace1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9e57f8f9-6e9e-45fb-91d1-132490e930ae', 'attached_at': '', 'detached_at': '', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'serial': '853747f9-2ca2-43a2-acfb-5d955470ace1'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 992.001937] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd19a605-4470-485b-a886-e6d618d6d1e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.025658] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c637fab-0356-49e4-b68d-73890d2ff17a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.033729] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b696e68-0d0a-43a9-a1fa-3a97dcda61ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.056475] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a65e8a-a727-4e62-9b4b-c4a9b4796fd7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.075061] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] The volume has not been displaced from its original location: [datastore1] volume-853747f9-2ca2-43a2-acfb-5d955470ace1/volume-853747f9-2ca2-43a2-acfb-5d955470ace1.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 992.081171] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Reconfiguring VM instance instance-00000034 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 992.081577] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06df96f5-60be-4608-8c1f-b1aa532e0611 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.106043] env[61986]: DEBUG oslo_vmware.api [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 992.106043] env[61986]: value = "task-1160217" [ 992.106043] env[61986]: _type = "Task" [ 992.106043] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.117772] env[61986]: DEBUG oslo_vmware.api [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.277351] env[61986]: INFO nova.compute.manager [-] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Took 1.37 seconds to deallocate network for instance. [ 992.283766] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160216, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.306033] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.306142] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.306403] env[61986]: DEBUG nova.compute.manager [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing instance network info cache due to event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 992.306729] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.373439] env[61986]: DEBUG nova.compute.manager [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Received event network-vif-plugged-9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.373736] env[61986]: DEBUG oslo_concurrency.lockutils [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] Acquiring lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.373927] env[61986]: DEBUG oslo_concurrency.lockutils [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.374136] env[61986]: DEBUG oslo_concurrency.lockutils [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.374329] env[61986]: DEBUG nova.compute.manager [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] No waiting events found dispatching network-vif-plugged-9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.374449] env[61986]: WARNING nova.compute.manager [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Received unexpected event network-vif-plugged-9f71edfb-58a5-43cb-9d35-62cda5bf87ad for instance with vm_state building and task_state spawning. [ 992.374615] env[61986]: DEBUG nova.compute.manager [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Received event network-changed-9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.374761] env[61986]: DEBUG nova.compute.manager [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Refreshing instance network info cache due to event network-changed-9f71edfb-58a5-43cb-9d35-62cda5bf87ad. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 992.374931] env[61986]: DEBUG oslo_concurrency.lockutils [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] Acquiring lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.375088] env[61986]: DEBUG oslo_concurrency.lockutils [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] Acquired lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.375266] env[61986]: DEBUG nova.network.neutron [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Refreshing network info cache for port 9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 992.407671] env[61986]: INFO nova.scheduler.client.report [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted allocation for migration f0c082cb-b5ab-49a7-a3ba-9675c6afa79d [ 992.449580] env[61986]: DEBUG oslo_concurrency.lockutils [None req-50f3f717-7f14-480d-9de5-d5271d021e9b tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "45d80bab-0140-4473-9dc2-4a5481838aed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.108s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.460303] env[61986]: WARNING nova.network.neutron [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] 44bde880-8646-44c4-8c3c-c1faf7d4a9ce already exists in list: networks containing: ['44bde880-8646-44c4-8c3c-c1faf7d4a9ce']. ignoring it [ 992.585134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462960df-fbf2-4977-a4ca-44fc6d23a814 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.592419] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7916e83-9731-4bbd-9905-f3da7e338c0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.623656] env[61986]: INFO nova.compute.manager [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Rebuilding instance [ 992.631197] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88387e7-2a00-456b-8b27-dcad48455158 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.642764] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75686a6c-d01a-477c-9b75-8115d3087181 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.648256] env[61986]: DEBUG oslo_vmware.api [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160217, 'name': ReconfigVM_Task, 'duration_secs': 0.246185} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.652904] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Reconfigured VM instance instance-00000034 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 992.660146] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21927446-bdce-455f-86c1-78623553e9a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.686094] env[61986]: DEBUG nova.compute.provider_tree [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.703714] env[61986]: DEBUG oslo_vmware.api [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 992.703714] env[61986]: value = "task-1160218" [ 992.703714] env[61986]: _type = "Task" [ 992.703714] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.704338] env[61986]: DEBUG nova.compute.manager [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.705325] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327d590e-e0c8-49b4-9294-73ada691b401 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.721464] env[61986]: DEBUG oslo_vmware.api [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160218, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.782407] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160216, 'name': CreateVM_Task, 'duration_secs': 0.528549} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.782755] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 992.783339] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.783517] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.783833] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.784109] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3f2a178-1d59-4427-b1b8-76305a535eaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.788924] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.789920] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 992.789920] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52502bc4-4f94-63b0-ba90-5c2b49e39e35" [ 992.789920] env[61986]: _type = "Task" [ 992.789920] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.798107] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52502bc4-4f94-63b0-ba90-5c2b49e39e35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.868751] env[61986]: DEBUG nova.network.neutron [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7228964a-4136-4e76-8fe4-a21ab14deacf", "address": "fa:16:3e:29:eb:9d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7228964a-41", "ovs_interfaceid": "7228964a-4136-4e76-8fe4-a21ab14deacf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.915275] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b4b39829-3e6c-4ae5-96e9-a9697c0c1aac tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.982s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.153936] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.154238] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.154459] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.154647] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.154817] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.157597] env[61986]: INFO nova.compute.manager [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Terminating instance [ 993.159091] env[61986]: DEBUG nova.compute.manager [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.159291] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 993.160155] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696b4574-be63-4ad7-8661-69ef3be451a6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.169087] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.169325] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a59e86b-4ef5-4786-87a6-4bb07e84297b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.173613] env[61986]: DEBUG nova.network.neutron [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updated VIF entry in instance network info cache for port 9f71edfb-58a5-43cb-9d35-62cda5bf87ad. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.173946] env[61986]: DEBUG nova.network.neutron [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance_info_cache with network_info: [{"id": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "address": "fa:16:3e:c3:f2:46", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f71edfb-58", "ovs_interfaceid": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.177419] env[61986]: DEBUG oslo_vmware.api [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 993.177419] env[61986]: value = "task-1160219" [ 993.177419] env[61986]: _type = "Task" [ 993.177419] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.186665] env[61986]: DEBUG oslo_vmware.api [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160219, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.191843] env[61986]: DEBUG nova.scheduler.client.report [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.195250] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.216484] env[61986]: DEBUG oslo_vmware.api [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160218, 'name': ReconfigVM_Task, 'duration_secs': 0.147863} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.216484] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252367', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'name': 'volume-853747f9-2ca2-43a2-acfb-5d955470ace1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9e57f8f9-6e9e-45fb-91d1-132490e930ae', 'attached_at': '', 'detached_at': '', 'volume_id': '853747f9-2ca2-43a2-acfb-5d955470ace1', 'serial': '853747f9-2ca2-43a2-acfb-5d955470ace1'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 993.227030] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.227221] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f895ffe8-d4cd-4a51-89d3-9a8f39dc4bbb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.234897] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 993.234897] env[61986]: value = "task-1160220" [ 993.234897] env[61986]: _type = "Task" [ 993.234897] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.244673] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.301147] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52502bc4-4f94-63b0-ba90-5c2b49e39e35, 'name': SearchDatastore_Task, 'duration_secs': 0.010888} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.301559] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.301724] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.301977] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.302146] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.302351] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.302633] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-545aede2-f5fc-442d-b2c7-581626c1b886 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.312991] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.313226] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 993.314081] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a03cd7c1-d4c7-440f-8931-41b365225d9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.320992] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 993.320992] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c42ff6-0df6-7f84-f2cd-4cd22e206245" [ 993.320992] env[61986]: _type = "Task" [ 993.320992] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.329912] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c42ff6-0df6-7f84-f2cd-4cd22e206245, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.372216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.373021] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.373223] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.373542] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.373841] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.375790] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2439c5af-668c-48e9-bb28-0a179ca1031e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.395818] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.396114] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.396302] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.396462] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.396631] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.396885] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.397144] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.397349] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.397552] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.397755] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.397947] env[61986]: DEBUG nova.virt.hardware [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.405297] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Reconfiguring VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 993.406779] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-350a1116-b862-48ee-bc77-c3923b6d22d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.426582] env[61986]: DEBUG oslo_vmware.api [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 993.426582] env[61986]: value = "task-1160221" [ 993.426582] env[61986]: _type = "Task" [ 993.426582] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.437229] env[61986]: DEBUG oslo_vmware.api [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160221, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.676757] env[61986]: DEBUG oslo_concurrency.lockutils [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] Releasing lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.677031] env[61986]: DEBUG nova.compute.manager [req-f344f689-7ad0-4b94-8489-41b1f50bd74e req-0af9cae2-8985-44a0-9a50-32098e9160fc service nova] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Received event network-vif-deleted-9afcb8ca-5b37-41f4-a4de-a97cf229512b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.688276] env[61986]: DEBUG oslo_vmware.api [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160219, 'name': PowerOffVM_Task, 'duration_secs': 0.238883} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.688572] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 993.688745] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 993.688997] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32f21fd6-6e7c-40ce-b950-118ee8bbcd0c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.696817] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.699198] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.637s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.717664] env[61986]: INFO nova.scheduler.client.report [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Deleted allocations for instance 5170aa51-3307-42b1-b0dd-645dd4036e5b [ 993.748542] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160220, 'name': PowerOffVM_Task, 'duration_secs': 0.16034} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.748718] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 993.748874] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 993.749784] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812961ee-dca2-4563-9a62-921b957e1df4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.758140] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 993.758531] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f073882-e72b-470a-bbc3-b89072cf9047 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.766278] env[61986]: DEBUG nova.objects.instance [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'flavor' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.768803] env[61986]: DEBUG nova.objects.instance [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'flavor' on Instance uuid 9e57f8f9-6e9e-45fb-91d1-132490e930ae {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.789374] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 993.789782] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 993.789782] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleting the datastore file [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.790089] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af9965b5-d188-45b0-9744-c7c237ac0276 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.798374] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 993.798374] env[61986]: value = "task-1160224" [ 993.798374] env[61986]: _type = "Task" [ 993.798374] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.804692] env[61986]: DEBUG nova.compute.manager [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-changed-7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 993.805034] env[61986]: DEBUG nova.compute.manager [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing instance network info cache due to event network-changed-7228964a-4136-4e76-8fe4-a21ab14deacf. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 993.805177] env[61986]: DEBUG oslo_concurrency.lockutils [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.809527] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.832370] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c42ff6-0df6-7f84-f2cd-4cd22e206245, 'name': SearchDatastore_Task, 'duration_secs': 0.01131} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.832657] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b3cedd-f12c-49a6-b510-ec4a9e7b5f85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.839184] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 993.839184] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524f1f9e-2aca-30f6-f641-4efa82afd771" [ 993.839184] env[61986]: _type = "Task" [ 993.839184] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.840123] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 993.840335] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 993.840523] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Deleting the datastore file [datastore1] 92077300-2e56-4dec-9e8d-9b566ac3fe0e {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.843770] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3608fdd-9462-49eb-95e5-cb42caea3e6b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.854668] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524f1f9e-2aca-30f6-f641-4efa82afd771, 'name': SearchDatastore_Task, 'duration_secs': 0.011591} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.856168] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.856433] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07/eb10ea87-8019-4f3b-af3f-32a901ec6a07.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 993.856879] env[61986]: DEBUG oslo_vmware.api [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 993.856879] env[61986]: value = "task-1160225" [ 993.856879] env[61986]: _type = "Task" [ 993.856879] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.857174] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f29fcec-d145-4500-a073-a7e5d3965e47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.867954] env[61986]: DEBUG oslo_vmware.api [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160225, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.869335] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 993.869335] env[61986]: value = "task-1160226" [ 993.869335] env[61986]: _type = "Task" [ 993.869335] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.877756] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160226, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.938027] env[61986]: DEBUG oslo_vmware.api [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.227900] env[61986]: DEBUG oslo_concurrency.lockutils [None req-24519c47-eb1f-4475-9aae-1c6d44b26102 tempest-ServersTestBootFromVolume-25130141 tempest-ServersTestBootFromVolume-25130141-project-member] Lock "5170aa51-3307-42b1-b0dd-645dd4036e5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.556s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.270482] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated VIF entry in instance network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 994.271328] env[61986]: DEBUG nova.network.neutron [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7228964a-4136-4e76-8fe4-a21ab14deacf", "address": "fa:16:3e:29:eb:9d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7228964a-41", "ovs_interfaceid": "7228964a-4136-4e76-8fe4-a21ab14deacf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.275738] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.275910] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.276127] env[61986]: DEBUG nova.network.neutron [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.276270] env[61986]: DEBUG nova.objects.instance [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'info_cache' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.309923] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101238} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.310255] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.312079] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.312079] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.370791] env[61986]: DEBUG oslo_vmware.api [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160225, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16262} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.376616] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.377158] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.377473] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.377849] env[61986]: INFO nova.compute.manager [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Took 1.22 seconds to destroy the instance on the hypervisor. [ 994.378270] env[61986]: DEBUG oslo.service.loopingcall [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.378935] env[61986]: DEBUG nova.compute.manager [-] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.379160] env[61986]: DEBUG nova.network.neutron [-] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.387261] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160226, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.440610] env[61986]: DEBUG oslo_vmware.api [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160221, 'name': ReconfigVM_Task, 'duration_secs': 0.696356} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.443582] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.443991] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Reconfigured VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 994.747022] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa12919e-3891-4da9-a280-4155137864e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.747186] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.747311] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9e57f8f9-6e9e-45fb-91d1-132490e930ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.747442] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance b9706e51-0f74-4dbc-aab1-ea640ac78c62 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 994.747560] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 59cfee64-ef79-4b8f-a703-c8812551d12f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.747682] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6488f68e-7b84-4462-aef5-25d02db504f6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 994.747797] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance bc93bdd5-256e-4900-a9ae-609b0a7d402d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.747911] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 221e10a3-da31-410c-80f8-4bcc2c515710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.748124] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance with task_state "deleting" is not being actively managed by this compute host but has allocations referencing this compute node (2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 994.748230] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.748320] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance bda7750f-eec9-40d9-ace5-18d48234126e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.748403] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 92077300-2e56-4dec-9e8d-9b566ac3fe0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.748509] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 73b2c463-0460-4a68-b788-ab0c7818d8f8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 994.748623] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 45d80bab-0140-4473-9dc2-4a5481838aed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.748735] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 8a594fe5-6fff-48ab-9f7f-474b2a24a486 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.748845] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance eb10ea87-8019-4f3b-af3f-32a901ec6a07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 994.749080] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 994.750299] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 994.751027] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 994.779745] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5d98994-8c1c-4f4d-b8e8-915354f23541 req-34286f7a-0bd6-42ae-97f1-a1bebf6c96b6 service nova] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.781600] env[61986]: DEBUG nova.objects.base [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Object Instance<92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f> lazy-loaded attributes: flavor,info_cache {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 994.785389] env[61986]: DEBUG oslo_concurrency.lockutils [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.785574] env[61986]: DEBUG nova.network.neutron [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing network info cache for port 7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 994.791254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-362e116b-0433-4696-ae6d-43b1fb01071f tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.329s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.885028] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160226, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540769} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.885028] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07/eb10ea87-8019-4f3b-af3f-32a901ec6a07.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 994.885028] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.885028] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93b5c26c-f075-4442-be70-c379c31c22ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.893106] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 994.893106] env[61986]: value = "task-1160227" [ 994.893106] env[61986]: _type = "Task" [ 994.893106] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.908592] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160227, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.949272] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5131ac34-0cc6-4eec-9b64-7eb55a417ee4 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bc93bdd5-256e-4900-a9ae-609b0a7d402d-7228964a-4136-4e76-8fe4-a21ab14deacf" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.610s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.043583] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842eaacc-df46-4347-bba2-321b745a8bcd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.053489] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f1b62b-996f-433f-acd8-79178011a6ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.092188] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f51a37-c509-4ed6-a1e3-977d47ad2d12 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.101570] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5068779c-ed42-468d-a1e4-22d84ab7cefd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.119099] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 995.357508] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.357821] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.357989] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.358199] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.358352] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.358497] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.358708] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.358871] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.359051] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.359219] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.359397] env[61986]: DEBUG nova.virt.hardware [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.361031] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26496a3-7f79-44d0-a147-e0fb913e910e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.372439] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b079ec94-03e2-4c69-b1b4-039cc7852931 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.388033] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.393784] env[61986]: DEBUG oslo.service.loopingcall [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.396458] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.396724] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2dc0fbd8-d7dc-4493-8be3-a0d68fdb034c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.418158] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160227, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074376} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.419587] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.419897] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.419897] env[61986]: value = "task-1160228" [ 995.419897] env[61986]: _type = "Task" [ 995.419897] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.420647] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bc9ed3-f983-471f-b1a5-5b42ae6d5542 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.432799] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160228, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.450579] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07/eb10ea87-8019-4f3b-af3f-32a901ec6a07.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.455011] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4316bc8-05c1-490f-8885-674b7d077c01 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.477017] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 995.477017] env[61986]: value = "task-1160229" [ 995.477017] env[61986]: _type = "Task" [ 995.477017] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.490118] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160229, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.642090] env[61986]: ERROR nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [req-54b7ab9d-92e0-41ac-992d-eed2ed07fbec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-54b7ab9d-92e0-41ac-992d-eed2ed07fbec"}]} [ 995.649245] env[61986]: DEBUG nova.network.neutron [-] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.653919] env[61986]: DEBUG nova.network.neutron [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [{"id": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "address": "fa:16:3e:c9:37:3f", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bca3360-2d", "ovs_interfaceid": "8bca3360-2dd5-4113-85c7-5e2e44864b07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.663054] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Refreshing inventories for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 995.674052] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.675497] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.676036] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.676338] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.676761] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.679399] env[61986]: INFO nova.compute.manager [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Terminating instance [ 995.681929] env[61986]: DEBUG nova.compute.manager [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.682197] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 995.683261] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643ae949-29e9-4262-ab3f-9e508b9f2a4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.687041] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updating ProviderTree inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 995.687212] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 162, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 995.698579] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 995.699017] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc853ea2-b62a-4e05-983b-c99c85aabb28 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.704781] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Refreshing aggregate associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, aggregates: None {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 995.708808] env[61986]: DEBUG oslo_vmware.api [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 995.708808] env[61986]: value = "task-1160230" [ 995.708808] env[61986]: _type = "Task" [ 995.708808] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.718142] env[61986]: DEBUG oslo_vmware.api [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160230, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.719229] env[61986]: DEBUG nova.network.neutron [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated VIF entry in instance network info cache for port 7228964a-4136-4e76-8fe4-a21ab14deacf. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 995.720183] env[61986]: DEBUG nova.network.neutron [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7228964a-4136-4e76-8fe4-a21ab14deacf", "address": "fa:16:3e:29:eb:9d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7228964a-41", "ovs_interfaceid": "7228964a-4136-4e76-8fe4-a21ab14deacf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.735139] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Refreshing trait associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 995.938891] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160228, 'name': CreateVM_Task, 'duration_secs': 0.50084} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.942099] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.942733] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.943289] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.943346] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.943599] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f447379-fc1f-447a-a7b4-29182c04452c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.949060] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 995.949060] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5216f1f4-4752-f5ac-b071-e1284f02eb82" [ 995.949060] env[61986]: _type = "Task" [ 995.949060] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.962638] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5216f1f4-4752-f5ac-b071-e1284f02eb82, 'name': SearchDatastore_Task, 'duration_secs': 0.009335} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.962638] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.962638] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.962885] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.962885] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.963061] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.963437] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99c5566a-4911-4c48-9c14-22a56fa24175 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.975026] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.975234] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.975971] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6d29a83-f71b-4eab-9763-933a4a47ba5d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.990942] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160229, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.994496] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 995.994496] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d34e95-bd60-2fba-94b3-d27ee2a48679" [ 995.994496] env[61986]: _type = "Task" [ 995.994496] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.005477] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d34e95-bd60-2fba-94b3-d27ee2a48679, 'name': SearchDatastore_Task, 'duration_secs': 0.009037} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.007770] env[61986]: DEBUG nova.compute.manager [req-0287ceb7-06cf-49d2-b3ff-98a731650c73 req-14ad5ced-e788-4870-8087-d0c0983155b7 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-vif-deleted-759ea478-47fb-45dc-a788-f500f86de7fd {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.007770] env[61986]: DEBUG nova.compute.manager [req-0287ceb7-06cf-49d2-b3ff-98a731650c73 req-14ad5ced-e788-4870-8087-d0c0983155b7 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-vif-deleted-0684bb22-0a77-4633-b834-a66fe18747c5 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.007770] env[61986]: DEBUG nova.compute.manager [req-0287ceb7-06cf-49d2-b3ff-98a731650c73 req-14ad5ced-e788-4870-8087-d0c0983155b7 service nova] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Received event network-vif-deleted-b3205584-42e2-4cb4-8b2d-0dd929934176 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.014731] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ee9722c-5ea0-416e-a6c1-1b195553ef9b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.022654] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 996.022654] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525b1faa-f6e4-26ab-4252-38c09eacb277" [ 996.022654] env[61986]: _type = "Task" [ 996.022654] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.035032] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525b1faa-f6e4-26ab-4252-38c09eacb277, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.045277] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55866053-bfa5-4cfa-87f1-6e134b7a9b0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.053657] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a86828-37fe-4f3c-be31-06b0629be92e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.091331] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539115c2-0efb-4a23-960c-2a9759f96bed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.102542] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e1b9b7-a402-441f-a53e-56c6d8e9db59 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.117496] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.151103] env[61986]: INFO nova.compute.manager [-] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Took 1.77 seconds to deallocate network for instance. [ 996.156490] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.219496] env[61986]: DEBUG oslo_vmware.api [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160230, 'name': PowerOffVM_Task, 'duration_secs': 0.232583} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.220403] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 996.220675] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 996.220983] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b099ff7-abac-4ed1-91c0-9396e719aabb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.223009] env[61986]: DEBUG oslo_concurrency.lockutils [req-1ef3d7ee-19cb-46b6-b873-608b7b673887 req-b82c848a-282e-4305-b5d2-1084fbfd0dc4 service nova] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.289673] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 996.289918] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 996.290223] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleting the datastore file [datastore2] 9e57f8f9-6e9e-45fb-91d1-132490e930ae {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.290512] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49237f27-7ba4-4963-8a6f-a07cfc730bae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.298289] env[61986]: DEBUG oslo_vmware.api [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 996.298289] env[61986]: value = "task-1160232" [ 996.298289] env[61986]: _type = "Task" [ 996.298289] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.310299] env[61986]: DEBUG oslo_vmware.api [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160232, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.493946] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160229, 'name': ReconfigVM_Task, 'duration_secs': 0.610572} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.494487] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Reconfigured VM instance instance-00000059 to attach disk [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07/eb10ea87-8019-4f3b-af3f-32a901ec6a07.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.495407] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d55b5ca4-626b-41fd-aac6-79ea49df8a37 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.503636] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 996.503636] env[61986]: value = "task-1160233" [ 996.503636] env[61986]: _type = "Task" [ 996.503636] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.515294] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160233, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.531029] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525b1faa-f6e4-26ab-4252-38c09eacb277, 'name': SearchDatastore_Task, 'duration_secs': 0.009412} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.531297] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.531743] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 996.531925] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-240e8b0b-02e6-4c8f-ad03-afb78fc19586 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.538098] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 996.538098] env[61986]: value = "task-1160234" [ 996.538098] env[61986]: _type = "Task" [ 996.538098] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.545837] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.659499] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.660415] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 996.660733] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-518b11c5-91e7-4ec1-896b-f45ad4d15208 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.668256] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updated inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with generation 113 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 996.668509] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updating resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d generation from 113 to 114 during operation: update_inventory {{(pid=61986) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 996.668673] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.673573] env[61986]: DEBUG oslo_vmware.api [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 996.673573] env[61986]: value = "task-1160235" [ 996.673573] env[61986]: _type = "Task" [ 996.673573] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.682900] env[61986]: DEBUG oslo_vmware.api [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160235, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.708591] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-bc93bdd5-256e-4900-a9ae-609b0a7d402d-7228964a-4136-4e76-8fe4-a21ab14deacf" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.708881] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bc93bdd5-256e-4900-a9ae-609b0a7d402d-7228964a-4136-4e76-8fe4-a21ab14deacf" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.813963] env[61986]: DEBUG oslo_vmware.api [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160232, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16569} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.817590] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.817590] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 996.817590] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 996.817590] env[61986]: INFO nova.compute.manager [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Took 1.13 seconds to destroy the instance on the hypervisor. [ 996.817590] env[61986]: DEBUG oslo.service.loopingcall [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.817590] env[61986]: DEBUG nova.compute.manager [-] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.817590] env[61986]: DEBUG nova.network.neutron [-] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 997.015192] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160233, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.048544] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160234, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484103} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.048838] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.049077] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.049350] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd1fdc61-9052-4dcf-9c53-865cb8f1c65f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.057559] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 997.057559] env[61986]: value = "task-1160236" [ 997.057559] env[61986]: _type = "Task" [ 997.057559] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.067095] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160236, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.175301] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 997.175745] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.477s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.176266] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.008s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.176588] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.183087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.993s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.183087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.183087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.584s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.183087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.185310] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.987s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.185684] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.189862] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.401s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.190242] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.192352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.533s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.192694] env[61986]: DEBUG nova.objects.instance [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lazy-loading 'resources' on Instance uuid 92077300-2e56-4dec-9e8d-9b566ac3fe0e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.205108] env[61986]: DEBUG oslo_vmware.api [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160235, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.215025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.215025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.215025] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad951d7e-88c9-4188-aad3-359d22f9ec47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.240282] env[61986]: INFO nova.scheduler.client.report [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted allocations for instance b9706e51-0f74-4dbc-aab1-ea640ac78c62 [ 997.247531] env[61986]: INFO nova.scheduler.client.report [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Deleted allocations for instance 3b925174-cf0b-4c6c-b6b6-1c99a50026be [ 997.247531] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b13336-f8be-4c61-a930-d35a65929edc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.259955] env[61986]: INFO nova.scheduler.client.report [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Deleted allocations for instance 6488f68e-7b84-4462-aef5-25d02db504f6 [ 997.290138] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Reconfiguring VM to detach interface {{(pid=61986) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 997.290394] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-275109dc-5276-49f6-bb47-5fdaf7ad7ea1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.312053] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 997.312053] env[61986]: value = "task-1160237" [ 997.312053] env[61986]: _type = "Task" [ 997.312053] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.324772] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.326659] env[61986]: INFO nova.scheduler.client.report [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Deleted allocations for instance 73b2c463-0460-4a68-b788-ab0c7818d8f8 [ 997.515023] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160233, 'name': Rename_Task, 'duration_secs': 0.740922} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.515362] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 997.515658] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1263f315-bfbb-4bae-af08-39324a9ff016 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.524134] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 997.524134] env[61986]: value = "task-1160238" [ 997.524134] env[61986]: _type = "Task" [ 997.524134] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.538369] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160238, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.574169] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160236, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083709} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.574169] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.574169] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7489c864-8f67-492a-a7b5-27f6b5f6cb15 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.596234] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.596548] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92844b66-ef05-4def-b72e-496c8992ba6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.622063] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 997.622063] env[61986]: value = "task-1160239" [ 997.622063] env[61986]: _type = "Task" [ 997.622063] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.630922] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.687269] env[61986]: DEBUG oslo_vmware.api [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160235, 'name': PowerOnVM_Task, 'duration_secs': 0.550946} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.688041] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 997.688041] env[61986]: DEBUG nova.compute.manager [None req-1aa8e3ac-22b3-4aa2-9d44-d190dbd45c05 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.688576] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325d8eaa-1cc5-49a5-9986-cad57e220834 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.708785] env[61986]: DEBUG oslo_concurrency.lockutils [None req-099ff59c-d723-48ed-bb36-000586f85f24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.737s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.709491] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.514s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.709684] env[61986]: INFO nova.compute.manager [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Unshelving [ 997.762934] env[61986]: DEBUG oslo_concurrency.lockutils [None req-08586ef4-56fc-480a-9b6f-f52ae5ac6a17 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "b9706e51-0f74-4dbc-aab1-ea640ac78c62" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.123s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.769694] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d404e1ac-48fc-4753-a988-31bb908c4e1a tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.589s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.773389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 12.237s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.773389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.773389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.773389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.773389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-77334961-b870-4ad1-853d-9eb557a5d3fc tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "6488f68e-7b84-4462-aef5-25d02db504f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.007s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.774732] env[61986]: INFO nova.compute.manager [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Terminating instance [ 997.778719] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquiring lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.778893] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Acquired lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.779093] env[61986]: DEBUG nova.network.neutron [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 997.824428] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.836351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-86bb817e-25f3-420b-8389-6ba479c2aea6 tempest-ImagesOneServerNegativeTestJSON-643238208 tempest-ImagesOneServerNegativeTestJSON-643238208-project-member] Lock "73b2c463-0460-4a68-b788-ab0c7818d8f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.315s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.918964] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e144da-a332-46c7-b419-a14663a5beaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.929382] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfd909a-f34f-4e2c-9c35-1405fd49628c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.970344] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaeb9de4-7423-402c-a31e-0fda5fb485e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.981794] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e92e5c-cdbe-42ec-9653-c1f692d91f18 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.998413] env[61986]: DEBUG nova.compute.provider_tree [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.038260] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160238, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.061377] env[61986]: DEBUG nova.compute.manager [req-b097bdb2-b07f-4656-93f5-57d400767539 req-e281c4ca-81f8-4f63-ab4d-13eba78b0aa3 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Received event network-vif-deleted-e570192b-e139-47dd-a44e-8f34242b8649 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.061679] env[61986]: INFO nova.compute.manager [req-b097bdb2-b07f-4656-93f5-57d400767539 req-e281c4ca-81f8-4f63-ab4d-13eba78b0aa3 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Neutron deleted interface e570192b-e139-47dd-a44e-8f34242b8649; detaching it from the instance and deleting it from the info cache [ 998.061734] env[61986]: DEBUG nova.network.neutron [req-b097bdb2-b07f-4656-93f5-57d400767539 req-e281c4ca-81f8-4f63-ab4d-13eba78b0aa3 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.134675] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.193033] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.193033] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.193033] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 998.217902] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "59cfee64-ef79-4b8f-a703-c8812551d12f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.217902] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.219078] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "59cfee64-ef79-4b8f-a703-c8812551d12f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.219351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.219647] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.226821] env[61986]: INFO nova.compute.manager [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Terminating instance [ 998.229131] env[61986]: DEBUG nova.compute.manager [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.229547] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 998.231114] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e4d81d-d0f1-4142-ae33-f681b1b31152 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.244277] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.244410] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18eadd5d-ad1e-4266-961d-21f94d95a0ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.251750] env[61986]: DEBUG oslo_vmware.api [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 998.251750] env[61986]: value = "task-1160240" [ 998.251750] env[61986]: _type = "Task" [ 998.251750] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.266989] env[61986]: DEBUG oslo_vmware.api [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.283606] env[61986]: DEBUG nova.compute.utils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Can not refresh info_cache because instance was not found {{(pid=61986) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 998.318664] env[61986]: DEBUG nova.network.neutron [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 998.327185] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.357126] env[61986]: DEBUG nova.network.neutron [-] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.505243] env[61986]: DEBUG nova.scheduler.client.report [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.539227] env[61986]: DEBUG oslo_vmware.api [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160238, 'name': PowerOnVM_Task, 'duration_secs': 1.010073} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.539549] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.541788] env[61986]: INFO nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Took 9.10 seconds to spawn the instance on the hypervisor. [ 998.542316] env[61986]: DEBUG nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.543932] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ec3c39-82a6-43de-aa01-789bddc4a0e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.568171] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1fd18bfa-88c3-4b3d-851b-7c90581421d2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.580821] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78128bf4-5668-4ec0-9735-7a3f1b96b7f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.621692] env[61986]: DEBUG nova.compute.manager [req-b097bdb2-b07f-4656-93f5-57d400767539 req-e281c4ca-81f8-4f63-ab4d-13eba78b0aa3 service nova] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Detach interface failed, port_id=e570192b-e139-47dd-a44e-8f34242b8649, reason: Instance 9e57f8f9-6e9e-45fb-91d1-132490e930ae could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 998.639554] env[61986]: DEBUG nova.network.neutron [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.641367] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160239, 'name': ReconfigVM_Task, 'duration_secs': 0.829901} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.641986] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486/8a594fe5-6fff-48ab-9f7f-474b2a24a486.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.642921] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f91ebf57-3038-42e3-805d-6b9eed7106cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.653163] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 998.653163] env[61986]: value = "task-1160241" [ 998.653163] env[61986]: _type = "Task" [ 998.653163] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.659476] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.659786] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.660156] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.660353] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.660564] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.663645] env[61986]: INFO nova.compute.manager [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Terminating instance [ 998.669065] env[61986]: DEBUG nova.compute.manager [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.669328] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 998.670668] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f03faa-4dad-4336-87c1-05432373b8ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.680627] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160241, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.683348] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.683643] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f7d798c-74f7-4b13-bf57-868ba2837306 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.693672] env[61986]: DEBUG oslo_vmware.api [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 998.693672] env[61986]: value = "task-1160242" [ 998.693672] env[61986]: _type = "Task" [ 998.693672] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.697415] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Didn't find any instances for network info cache update. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 998.697622] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.698244] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.698422] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.698577] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.698719] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.705924] env[61986]: DEBUG oslo_vmware.api [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.751218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.762039] env[61986]: DEBUG oslo_vmware.api [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160240, 'name': PowerOffVM_Task, 'duration_secs': 0.258214} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.762340] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 998.762543] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 998.762810] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e435db33-b26b-47ab-9a90-0fd1f6f82bbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.827173] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.841907] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 998.842163] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 998.842351] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Deleting the datastore file [datastore2] 59cfee64-ef79-4b8f-a703-c8812551d12f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.842677] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f92b3733-93a5-4f5b-b91e-3c23b9730802 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.855660] env[61986]: DEBUG oslo_vmware.api [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for the task: (returnval){ [ 998.855660] env[61986]: value = "task-1160244" [ 998.855660] env[61986]: _type = "Task" [ 998.855660] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.865037] env[61986]: INFO nova.compute.manager [-] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Took 2.05 seconds to deallocate network for instance. [ 998.877911] env[61986]: DEBUG oslo_vmware.api [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160244, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.012265] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.819s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.016306] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.264s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.016306] env[61986]: DEBUG nova.objects.instance [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lazy-loading 'pci_requests' on Instance uuid ee47cab9-5dfd-48ce-ba70-cb800d735b19 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.066272] env[61986]: INFO nova.scheduler.client.report [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Deleted allocations for instance 92077300-2e56-4dec-9e8d-9b566ac3fe0e [ 999.084869] env[61986]: INFO nova.compute.manager [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Took 28.17 seconds to build instance. [ 999.144901] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Releasing lock "refresh_cache-3b925174-cf0b-4c6c-b6b6-1c99a50026be" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.145392] env[61986]: DEBUG nova.compute.manager [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.145597] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 999.145896] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87fc5828-0317-447a-ae81-ab7aa416a2f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.163019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775a00e7-08cd-4c27-bd4c-dd8c4d0f12d2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.185513] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160241, 'name': Rename_Task, 'duration_secs': 0.21654} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.185513] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.185513] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9befc306-44e7-4cd8-b5ff-257889f81a75 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.193158] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 999.193158] env[61986]: value = "task-1160245" [ 999.193158] env[61986]: _type = "Task" [ 999.193158] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.204310] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3b925174-cf0b-4c6c-b6b6-1c99a50026be could not be found. [ 999.205037] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 999.205037] env[61986]: INFO nova.compute.manager [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Took 0.06 seconds to destroy the instance on the hypervisor. [ 999.205037] env[61986]: DEBUG oslo.service.loopingcall [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.211863] env[61986]: DEBUG nova.compute.manager [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.211863] env[61986]: DEBUG nova.network.neutron [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 999.229734] env[61986]: DEBUG oslo_vmware.api [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160242, 'name': PowerOffVM_Task, 'duration_secs': 0.261666} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.230046] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160245, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.230363] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.230523] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 999.230798] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c068a702-fcf0-46f3-8770-552d78699a03 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.265053] env[61986]: DEBUG nova.network.neutron [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 999.302878] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 999.303175] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 999.303459] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleting the datastore file [datastore2] aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.304449] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8313f605-c456-460d-9f6a-b0abd5e7a3e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.315598] env[61986]: DEBUG oslo_vmware.api [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 999.315598] env[61986]: value = "task-1160247" [ 999.315598] env[61986]: _type = "Task" [ 999.315598] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.336967] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.338115] env[61986]: DEBUG oslo_vmware.api [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.367473] env[61986]: DEBUG oslo_vmware.api [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Task: {'id': task-1160244, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186912} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.368041] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.368379] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 999.368738] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 999.369074] env[61986]: INFO nova.compute.manager [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 999.369485] env[61986]: DEBUG oslo.service.loopingcall [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.369836] env[61986]: DEBUG nova.compute.manager [-] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.370160] env[61986]: DEBUG nova.network.neutron [-] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 999.373212] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.503922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.504261] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.504477] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.504667] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.504845] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.507232] env[61986]: INFO nova.compute.manager [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Terminating instance [ 999.509559] env[61986]: DEBUG nova.compute.manager [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.509771] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 999.510710] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf99671-479d-4f5f-a83e-844c98a2a186 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.520782] env[61986]: DEBUG nova.objects.instance [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lazy-loading 'numa_topology' on Instance uuid ee47cab9-5dfd-48ce-ba70-cb800d735b19 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.522503] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.522782] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c3399ba-9abd-4319-914e-b2cb6b6920f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.531488] env[61986]: DEBUG oslo_vmware.api [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 999.531488] env[61986]: value = "task-1160248" [ 999.531488] env[61986]: _type = "Task" [ 999.531488] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.542212] env[61986]: DEBUG oslo_vmware.api [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160248, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.578149] env[61986]: DEBUG oslo_concurrency.lockutils [None req-52e11320-75b8-4c8c-8c85-a8ea1f090f5b tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "92077300-2e56-4dec-9e8d-9b566ac3fe0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.424s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.584943] env[61986]: DEBUG oslo_concurrency.lockutils [None req-03fb20cd-69a3-4ded-b245-9ea51e4f43a3 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.679s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.716323] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160245, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.767740] env[61986]: DEBUG nova.network.neutron [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.828959] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.835215] env[61986]: DEBUG oslo_vmware.api [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151685} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.835561] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.835788] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 999.836024] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 999.836253] env[61986]: INFO nova.compute.manager [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 999.836645] env[61986]: DEBUG oslo.service.loopingcall [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.836963] env[61986]: DEBUG nova.compute.manager [-] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.837047] env[61986]: DEBUG nova.network.neutron [-] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1000.023795] env[61986]: INFO nova.compute.claims [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.043219] env[61986]: DEBUG oslo_vmware.api [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160248, 'name': PowerOffVM_Task, 'duration_secs': 0.369396} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.043879] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.044068] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1000.044339] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d2a44ff-fe3f-4a2d-a9ef-7b34cf4a6c07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.105657] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1000.105991] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1000.105991] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleting the datastore file [datastore2] 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.106270] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55245960-1626-4ec9-bc8d-e477021db585 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.114501] env[61986]: DEBUG oslo_vmware.api [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1000.114501] env[61986]: value = "task-1160250" [ 1000.114501] env[61986]: _type = "Task" [ 1000.114501] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.125275] env[61986]: DEBUG oslo_vmware.api [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160250, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.221774] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160245, 'name': PowerOnVM_Task} progress is 79%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.270988] env[61986]: INFO nova.compute.manager [-] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Took 1.06 seconds to deallocate network for instance. [ 1000.331855] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.628249] env[61986]: DEBUG oslo_vmware.api [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160250, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32686} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.631498] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.631498] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1000.631498] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1000.631498] env[61986]: INFO nova.compute.manager [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1000.631498] env[61986]: DEBUG oslo.service.loopingcall [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.631498] env[61986]: DEBUG nova.network.neutron [-] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.631498] env[61986]: DEBUG nova.compute.manager [-] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.631498] env[61986]: DEBUG nova.network.neutron [-] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1000.648432] env[61986]: DEBUG nova.compute.manager [req-49de62c7-4f1f-4110-a283-92e1f707e0c4 req-eb26532b-b421-4f69-b3a5-7c7c4e94d89d service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Received event network-vif-deleted-d9a2fed5-6f97-43b2-9e7a-5c859c341ebe {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1000.648660] env[61986]: INFO nova.compute.manager [req-49de62c7-4f1f-4110-a283-92e1f707e0c4 req-eb26532b-b421-4f69-b3a5-7c7c4e94d89d service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Neutron deleted interface d9a2fed5-6f97-43b2-9e7a-5c859c341ebe; detaching it from the instance and deleting it from the info cache [ 1000.648786] env[61986]: DEBUG nova.network.neutron [req-49de62c7-4f1f-4110-a283-92e1f707e0c4 req-eb26532b-b421-4f69-b3a5-7c7c4e94d89d service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.717147] env[61986]: DEBUG oslo_vmware.api [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160245, 'name': PowerOnVM_Task, 'duration_secs': 1.327371} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.718738] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1000.718891] env[61986]: DEBUG nova.compute.manager [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.719814] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ed3b30-ebcc-48eb-91af-52a1913b998f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.755470] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8160235-5223-4e97-8271-23b76f168b15 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.766345] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e7bcbc-c507-4317-8346-c485bc209e5f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.774592] env[61986]: DEBUG nova.compute.manager [req-d2e66f1f-390f-4534-9266-7344463f0586 req-ad0faad5-8878-46b6-bbc0-8fc67c4da4f9 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Received event network-vif-deleted-7b26e9ba-74ad-493c-8508-422bdb97e8db {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1000.774820] env[61986]: INFO nova.compute.manager [req-d2e66f1f-390f-4534-9266-7344463f0586 req-ad0faad5-8878-46b6-bbc0-8fc67c4da4f9 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Neutron deleted interface 7b26e9ba-74ad-493c-8508-422bdb97e8db; detaching it from the instance and deleting it from the info cache [ 1000.774975] env[61986]: DEBUG nova.network.neutron [req-d2e66f1f-390f-4534-9266-7344463f0586 req-ad0faad5-8878-46b6-bbc0-8fc67c4da4f9 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.800791] env[61986]: INFO nova.compute.manager [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance disappeared during terminate [ 1000.801133] env[61986]: DEBUG oslo_concurrency.lockutils [None req-224f9192-44f8-4fa5-a976-d34e3f85b542 tempest-ImagesTestJSON-1910434850 tempest-ImagesTestJSON-1910434850-project-member] Lock "3b925174-cf0b-4c6c-b6b6-1c99a50026be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.030s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.807018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fca868-a0c9-445e-be05-ad3b1bcfea56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.808379] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e86bbeae-053d-44ef-a506-563afe1da733 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.816867] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00408f77-6f82-4a30-8392-c17d3e01911e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.828196] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46582f58-e3a2-4a82-b7e1-e77ce7cf7b98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.847939] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.856887] env[61986]: DEBUG nova.compute.provider_tree [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.865259] env[61986]: DEBUG nova.network.neutron [-] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.866894] env[61986]: DEBUG nova.compute.manager [req-d2e66f1f-390f-4534-9266-7344463f0586 req-ad0faad5-8878-46b6-bbc0-8fc67c4da4f9 service nova] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Detach interface failed, port_id=7b26e9ba-74ad-493c-8508-422bdb97e8db, reason: Instance 59cfee64-ef79-4b8f-a703-c8812551d12f could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1001.136387] env[61986]: INFO nova.compute.manager [-] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Took 1.30 seconds to deallocate network for instance. [ 1001.159914] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-777c2d70-9f23-480b-b4d3-c3efcc4943d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.176449] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6103b4-3767-4079-9963-0174dc2b8e7d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.222173] env[61986]: DEBUG nova.compute.manager [req-49de62c7-4f1f-4110-a283-92e1f707e0c4 req-eb26532b-b421-4f69-b3a5-7c7c4e94d89d service nova] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Detach interface failed, port_id=d9a2fed5-6f97-43b2-9e7a-5c859c341ebe, reason: Instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1001.249418] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.343065] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.348961] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.349186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.372399] env[61986]: DEBUG nova.scheduler.client.report [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.378662] env[61986]: INFO nova.compute.manager [-] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Took 2.01 seconds to deallocate network for instance. [ 1001.660891] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.665684] env[61986]: DEBUG nova.compute.manager [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Stashing vm_state: active {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1001.837833] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.852319] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.879815] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.865s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.882552] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.509s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.883148] env[61986]: DEBUG nova.objects.instance [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'resources' on Instance uuid 9e57f8f9-6e9e-45fb-91d1-132490e930ae {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.888993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.961508] env[61986]: INFO nova.network.neutron [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating port 78729401-474d-4e4f-8159-aff2e908be0b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1002.071191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.071529] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.071713] env[61986]: INFO nova.compute.manager [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Shelving [ 1002.148020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.148020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.148020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.148020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.148020] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.149694] env[61986]: INFO nova.compute.manager [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Terminating instance [ 1002.152496] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "refresh_cache-8a594fe5-6fff-48ab-9f7f-474b2a24a486" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.152655] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "refresh_cache-8a594fe5-6fff-48ab-9f7f-474b2a24a486" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.152820] env[61986]: DEBUG nova.network.neutron [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.194253] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.335883] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.379756] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.410169] env[61986]: DEBUG nova.network.neutron [-] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.581489] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1002.581824] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e9dc03c-5535-4a54-88c6-ea236b1ba5af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.591422] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1002.591422] env[61986]: value = "task-1160251" [ 1002.591422] env[61986]: _type = "Task" [ 1002.591422] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.601568] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.675075] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4767b7e3-ce3b-4dc4-a1ab-e5a65353902c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.684028] env[61986]: DEBUG nova.network.neutron [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1002.687143] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035bda6b-8816-46cb-9961-f04ad351b2e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.724733] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2e8c50-859d-4a80-9ff3-847244647dec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.733126] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd441be-7f83-4955-8cc2-24f31bc76e44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.749623] env[61986]: DEBUG nova.compute.provider_tree [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.788311] env[61986]: DEBUG nova.network.neutron [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.827459] env[61986]: DEBUG nova.compute.manager [req-bcba2071-54e1-4374-b219-5a404053521c req-b6cb17cc-081f-42ea-8af4-dc0f1003d3b1 service nova] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Received event network-vif-deleted-8bca3360-2dd5-4113-85c7-5e2e44864b07 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1002.840658] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.914363] env[61986]: INFO nova.compute.manager [-] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Took 2.28 seconds to deallocate network for instance. [ 1003.106513] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160251, 'name': PowerOffVM_Task, 'duration_secs': 0.205184} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.106847] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1003.107720] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6635059-53e0-41af-999e-81b2b231cd9e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.134186] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc15c30-db2d-4a40-91a0-cb01cd1b61bd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.255328] env[61986]: DEBUG nova.scheduler.client.report [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.292724] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "refresh_cache-8a594fe5-6fff-48ab-9f7f-474b2a24a486" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.293197] env[61986]: DEBUG nova.compute.manager [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.293398] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.294293] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9feed1cc-fa5b-4445-b2c9-87d575db8379 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.307485] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.307822] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-007805a0-6024-46ec-80ea-51d989ba40ff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.315616] env[61986]: DEBUG oslo_vmware.api [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 1003.315616] env[61986]: value = "task-1160252" [ 1003.315616] env[61986]: _type = "Task" [ 1003.315616] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.325313] env[61986]: DEBUG oslo_vmware.api [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.336395] env[61986]: DEBUG oslo_vmware.api [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160237, 'name': ReconfigVM_Task, 'duration_secs': 5.843783} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.336743] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.337040] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Reconfigured VM to detach interface {{(pid=61986) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1003.422742] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.648837] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1003.650425] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7107da3f-c1af-42d3-8cab-0adf47e38c10 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.658960] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1003.658960] env[61986]: value = "task-1160253" [ 1003.658960] env[61986]: _type = "Task" [ 1003.658960] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.672070] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160253, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.680022] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.680022] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.680022] env[61986]: DEBUG nova.network.neutron [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.758323] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.761110] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.511s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.761229] env[61986]: DEBUG nova.objects.instance [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1003.788297] env[61986]: INFO nova.scheduler.client.report [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted allocations for instance 9e57f8f9-6e9e-45fb-91d1-132490e930ae [ 1003.835280] env[61986]: DEBUG oslo_vmware.api [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160252, 'name': PowerOffVM_Task, 'duration_secs': 0.18402} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.835568] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1003.836014] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1003.836323] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc23f592-2acb-4a19-8749-32cd16e8ee11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.865513] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1003.865749] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1003.865943] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleting the datastore file [datastore1] 8a594fe5-6fff-48ab-9f7f-474b2a24a486 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.866364] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-638ff0c5-27d9-403a-9e3a-59cc635579dc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.874166] env[61986]: DEBUG oslo_vmware.api [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 1003.874166] env[61986]: value = "task-1160255" [ 1003.874166] env[61986]: _type = "Task" [ 1003.874166] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.884573] env[61986]: DEBUG oslo_vmware.api [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.054179] env[61986]: DEBUG nova.compute.manager [req-db2d9935-a547-450b-b4fa-2b20b6f58725 req-9bf84094-0582-496b-b901-d9690038a858 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-vif-plugged-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1004.054494] env[61986]: DEBUG oslo_concurrency.lockutils [req-db2d9935-a547-450b-b4fa-2b20b6f58725 req-9bf84094-0582-496b-b901-d9690038a858 service nova] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.054812] env[61986]: DEBUG oslo_concurrency.lockutils [req-db2d9935-a547-450b-b4fa-2b20b6f58725 req-9bf84094-0582-496b-b901-d9690038a858 service nova] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.055318] env[61986]: DEBUG oslo_concurrency.lockutils [req-db2d9935-a547-450b-b4fa-2b20b6f58725 req-9bf84094-0582-496b-b901-d9690038a858 service nova] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.055736] env[61986]: DEBUG nova.compute.manager [req-db2d9935-a547-450b-b4fa-2b20b6f58725 req-9bf84094-0582-496b-b901-d9690038a858 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] No waiting events found dispatching network-vif-plugged-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.055736] env[61986]: WARNING nova.compute.manager [req-db2d9935-a547-450b-b4fa-2b20b6f58725 req-9bf84094-0582-496b-b901-d9690038a858 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received unexpected event network-vif-plugged-78729401-474d-4e4f-8159-aff2e908be0b for instance with vm_state shelved_offloaded and task_state spawning. [ 1004.171676] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160253, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.296313] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8fd4c3c5-bb92-494d-b8cb-fb33aa81cfb2 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "9e57f8f9-6e9e-45fb-91d1-132490e930ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.621s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.399177] env[61986]: DEBUG oslo_vmware.api [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099561} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.400580] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.400580] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.400580] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.400580] env[61986]: INFO nova.compute.manager [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1004.400875] env[61986]: DEBUG oslo.service.loopingcall [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.402080] env[61986]: DEBUG nova.compute.manager [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.402080] env[61986]: DEBUG nova.network.neutron [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.423338] env[61986]: DEBUG nova.network.neutron [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1004.562439] env[61986]: DEBUG nova.network.neutron [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.674985] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160253, 'name': CreateSnapshot_Task, 'duration_secs': 0.86174} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.675846] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1004.676689] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd0c313-c3fe-42d0-845c-bd49c4fcf75a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.782316] env[61986]: DEBUG oslo_concurrency.lockutils [None req-dc34ae6a-3980-41a0-a66b-343dfe7c69c5 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.783434] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.123s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.783667] env[61986]: DEBUG nova.objects.instance [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lazy-loading 'resources' on Instance uuid aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.788923] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "645ecfee-3c0b-47c1-86c1-8ca21781438a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.789316] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.928492] env[61986]: DEBUG nova.network.neutron [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.067841] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.108926] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='21630f5f0a94d9529028d40d71e9afae',container_format='bare',created_at=2024-10-07T06:45:24Z,direct_url=,disk_format='vmdk',id=24ded09f-c9f0-435c-93fc-1aef721bbca3,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-663700515-shelved',owner='c94c168de68c4608b772cf6fa6012dcb',properties=ImageMetaProps,protected=,size=31666176,status='active',tags=,updated_at=2024-10-07T06:45:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.109249] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.109456] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.109694] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.109846] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.109998] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.110470] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.110783] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.111072] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.111484] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.111484] env[61986]: DEBUG nova.virt.hardware [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.112435] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676b8413-76e3-4057-8d35-2f5ea43fa2f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.122415] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15605894-f976-4d16-bec8-d89cd3b05b2d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.138667] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.138904] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.139347] env[61986]: DEBUG nova.network.neutron [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.140807] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:fd:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e350f83a-f581-4e10-ac16-0b0f7bfd3d38', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78729401-474d-4e4f-8159-aff2e908be0b', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.148446] env[61986]: DEBUG oslo.service.loopingcall [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.149500] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1005.149775] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6090d37c-41dc-4dc3-a392-27c4cf9db193 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.175724] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.175724] env[61986]: value = "task-1160256" [ 1005.175724] env[61986]: _type = "Task" [ 1005.175724] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.186711] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160256, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.200918] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1005.201399] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-92834ed2-4cf7-48c5-a5e6-062036581027 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.210883] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1005.210883] env[61986]: value = "task-1160257" [ 1005.210883] env[61986]: _type = "Task" [ 1005.210883] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.220695] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160257, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.293039] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.433808] env[61986]: INFO nova.compute.manager [-] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Took 1.03 seconds to deallocate network for instance. [ 1005.585640] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8c1782-8112-4c13-8780-ff1381ae8a2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.596173] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4391874a-6dd2-4f49-8d71-aac1ebaa0e3d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.633312] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d5eec1-0644-45f0-8ef9-a80ebf692662 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.643210] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638ce679-692c-491e-92fe-b506a60970d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.662583] env[61986]: DEBUG nova.compute.provider_tree [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.686205] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160256, 'name': CreateVM_Task, 'duration_secs': 0.478034} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.686389] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1005.687049] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.687235] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.687617] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.688165] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1749f08-5232-4215-af36-7e439f9abdb5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.694178] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1005.694178] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d70cbc-8652-48c8-fcf5-80b05427b09c" [ 1005.694178] env[61986]: _type = "Task" [ 1005.694178] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.702935] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d70cbc-8652-48c8-fcf5-80b05427b09c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.722866] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160257, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.818335] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.943121] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.092289] env[61986]: DEBUG nova.compute.manager [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-changed-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1006.092579] env[61986]: DEBUG nova.compute.manager [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Refreshing instance network info cache due to event network-changed-78729401-474d-4e4f-8159-aff2e908be0b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1006.092899] env[61986]: DEBUG oslo_concurrency.lockutils [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.093123] env[61986]: DEBUG oslo_concurrency.lockutils [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.093345] env[61986]: DEBUG nova.network.neutron [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Refreshing network info cache for port 78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1006.160987] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "68626872-0875-4ff1-81ad-85bdff30ac94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.161241] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "68626872-0875-4ff1-81ad-85bdff30ac94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.168065] env[61986]: DEBUG nova.scheduler.client.report [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.208698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.209085] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Processing image 24ded09f-c9f0-435c-93fc-1aef721bbca3 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.209393] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.209603] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.209847] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.210204] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab5fb395-e891-4a7c-b730-7194d63a7f0f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.225166] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160257, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.234548] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.234776] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1006.235599] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e2d88cf-38f7-4c36-8c1d-0f1ba0de7eb2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.243356] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1006.243356] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c4c9e0-7143-7f18-7a3d-2d9021f547d2" [ 1006.243356] env[61986]: _type = "Task" [ 1006.243356] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.254275] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c4c9e0-7143-7f18-7a3d-2d9021f547d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.539556] env[61986]: INFO nova.network.neutron [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Port 7228964a-4136-4e76-8fe4-a21ab14deacf from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1006.539945] env[61986]: DEBUG nova.network.neutron [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.664377] env[61986]: DEBUG nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.673364] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.676263] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.786s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.676263] env[61986]: DEBUG nova.objects.instance [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lazy-loading 'resources' on Instance uuid 59cfee64-ef79-4b8f-a703-c8812551d12f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.719131] env[61986]: INFO nova.scheduler.client.report [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted allocations for instance aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d [ 1006.733448] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160257, 'name': CloneVM_Task, 'duration_secs': 1.443488} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.735549] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Created linked-clone VM from snapshot [ 1006.735549] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973d1d48-24cf-42d2-b1af-afa394139a45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.745754] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Uploading image 15c86942-389c-4a86-952d-179d0c1d0572 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1006.765259] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1006.765674] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Fetch image to [datastore2] OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b/OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1006.765863] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Downloading stream optimized image 24ded09f-c9f0-435c-93fc-1aef721bbca3 to [datastore2] OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b/OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b.vmdk on the data store datastore2 as vApp {{(pid=61986) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1006.766046] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Downloading image file data 24ded09f-c9f0-435c-93fc-1aef721bbca3 to the ESX as VM named 'OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b' {{(pid=61986) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1006.784123] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1006.784123] env[61986]: value = "vm-252436" [ 1006.784123] env[61986]: _type = "VirtualMachine" [ 1006.784123] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1006.784587] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-690b66c5-96db-4ca7-9a2d-77ca7a47d27e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.795603] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease: (returnval){ [ 1006.795603] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5265f84a-1b0a-a709-6160-9d67adca59ee" [ 1006.795603] env[61986]: _type = "HttpNfcLease" [ 1006.795603] env[61986]: } obtained for exporting VM: (result){ [ 1006.795603] env[61986]: value = "vm-252436" [ 1006.795603] env[61986]: _type = "VirtualMachine" [ 1006.795603] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1006.795912] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the lease: (returnval){ [ 1006.795912] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5265f84a-1b0a-a709-6160-9d67adca59ee" [ 1006.795912] env[61986]: _type = "HttpNfcLease" [ 1006.795912] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1006.819962] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.819962] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5265f84a-1b0a-a709-6160-9d67adca59ee" [ 1006.819962] env[61986]: _type = "HttpNfcLease" [ 1006.819962] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1006.820580] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1006.820580] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5265f84a-1b0a-a709-6160-9d67adca59ee" [ 1006.820580] env[61986]: _type = "HttpNfcLease" [ 1006.820580] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1006.821031] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0c9038-f57a-4fc1-bb4f-1b2fe70038a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.832819] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524967b0-4fdf-cd40-485b-5e11ee0b1d7c/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1006.833014] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524967b0-4fdf-cd40-485b-5e11ee0b1d7c/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1006.899032] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1006.899032] env[61986]: value = "resgroup-9" [ 1006.899032] env[61986]: _type = "ResourcePool" [ 1006.899032] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1006.899032] env[61986]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-6951f6c8-6c4e-4e21-b6f4-7829e3f955da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.920883] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lease: (returnval){ [ 1006.920883] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e7cef-4824-5ffb-00d1-da957c97ae42" [ 1006.920883] env[61986]: _type = "HttpNfcLease" [ 1006.920883] env[61986]: } obtained for vApp import into resource pool (val){ [ 1006.920883] env[61986]: value = "resgroup-9" [ 1006.920883] env[61986]: _type = "ResourcePool" [ 1006.920883] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1006.921264] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the lease: (returnval){ [ 1006.921264] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e7cef-4824-5ffb-00d1-da957c97ae42" [ 1006.921264] env[61986]: _type = "HttpNfcLease" [ 1006.921264] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1006.929300] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.929300] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e7cef-4824-5ffb-00d1-da957c97ae42" [ 1006.929300] env[61986]: _type = "HttpNfcLease" [ 1006.929300] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1006.947077] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b9ee0a0d-7587-4b8b-ab74-5e2e855d34b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.043666] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.174937] env[61986]: DEBUG nova.network.neutron [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updated VIF entry in instance network info cache for port 78729401-474d-4e4f-8159-aff2e908be0b. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1007.175723] env[61986]: DEBUG nova.network.neutron [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.201036] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.239200] env[61986]: DEBUG oslo_concurrency.lockutils [None req-19fc5f30-0e7d-4914-81d1-f9bb776fb3cc tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.579s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.337154] env[61986]: DEBUG nova.compute.manager [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.337835] env[61986]: DEBUG nova.compute.manager [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing instance network info cache due to event network-changed-51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1007.337938] env[61986]: DEBUG oslo_concurrency.lockutils [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.338463] env[61986]: DEBUG oslo_concurrency.lockutils [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.338463] env[61986]: DEBUG nova.network.neutron [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Refreshing network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1007.433265] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1007.433265] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e7cef-4824-5ffb-00d1-da957c97ae42" [ 1007.433265] env[61986]: _type = "HttpNfcLease" [ 1007.433265] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1007.445018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23796bc2-f01c-4a55-a203-82d7333720c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.456292] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5be177-c7db-44e9-a63c-cd8a6e4ef48b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.489911] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2dc3574-f5d1-4427-b11d-46aab0591462 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.499372] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8e4a99-3ad8-4214-8217-114b562ce685 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.518727] env[61986]: DEBUG nova.compute.provider_tree [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.549922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-156428bc-1c3f-400a-afad-c0459839233d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bc93bdd5-256e-4900-a9ae-609b0a7d402d-7228964a-4136-4e76-8fe4-a21ab14deacf" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.841s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.571057] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.571689] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.679942] env[61986]: DEBUG oslo_concurrency.lockutils [req-a3185d9b-a0da-4985-b610-b861e4f87399 req-3374952d-86ba-42c7-b7f8-0fd2a6b042e1 service nova] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.923157] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-bda7750f-eec9-40d9-ace5-18d48234126e-7228964a-4136-4e76-8fe4-a21ab14deacf" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.923157] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bda7750f-eec9-40d9-ace5-18d48234126e-7228964a-4136-4e76-8fe4-a21ab14deacf" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.923485] env[61986]: DEBUG nova.objects.instance [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'flavor' on Instance uuid bda7750f-eec9-40d9-ace5-18d48234126e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.937535] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1007.937535] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e7cef-4824-5ffb-00d1-da957c97ae42" [ 1007.937535] env[61986]: _type = "HttpNfcLease" [ 1007.937535] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1007.940126] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1007.940126] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521e7cef-4824-5ffb-00d1-da957c97ae42" [ 1007.940126] env[61986]: _type = "HttpNfcLease" [ 1007.940126] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1007.941743] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58679c53-efb7-4999-9dae-c70562dbb320 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.951976] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529ddd5a-8c89-4ba8-cdce-9c67ecc1843e/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1007.952271] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating HTTP connection to write to file with size = 31666176 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529ddd5a-8c89-4ba8-cdce-9c67ecc1843e/disk-0.vmdk. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1008.025493] env[61986]: DEBUG nova.scheduler.client.report [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.035121] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e9ccd396-a494-4578-b782-c4447209013d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.074310] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.168974] env[61986]: DEBUG nova.network.neutron [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated VIF entry in instance network info cache for port 51b89bbc-3efd-4e51-b21b-bfce45546eae. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1008.169767] env[61986]: DEBUG nova.network.neutron [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [{"id": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "address": "fa:16:3e:10:b1:97", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b89bbc-3e", "ovs_interfaceid": "51b89bbc-3efd-4e51-b21b-bfce45546eae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.531504] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.535434] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.341s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.547383] env[61986]: DEBUG nova.objects.instance [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'pci_requests' on Instance uuid bda7750f-eec9-40d9-ace5-18d48234126e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.564925] env[61986]: INFO nova.scheduler.client.report [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Deleted allocations for instance 59cfee64-ef79-4b8f-a703-c8812551d12f [ 1008.599386] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.674135] env[61986]: DEBUG oslo_concurrency.lockutils [req-96f402fd-0268-4881-8b54-0155d6646fb0 req-7448230a-5546-4c7d-9adc-d7fb92284a83 service nova] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.041125] env[61986]: INFO nova.compute.claims [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.050080] env[61986]: DEBUG nova.objects.base [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1009.050410] env[61986]: DEBUG nova.network.neutron [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1009.079111] env[61986]: DEBUG oslo_concurrency.lockutils [None req-85b61d45-3a2d-493a-b5b6-431166be6c42 tempest-ListImageFiltersTestJSON-1901603042 tempest-ListImageFiltersTestJSON-1901603042-project-member] Lock "59cfee64-ef79-4b8f-a703-c8812551d12f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.861s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.145553] env[61986]: DEBUG nova.policy [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85e78188137c42faaa50ba3e3a2bffda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '617d3b641cdc4d1b815b5edf7e970d62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1009.287206] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1009.287492] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529ddd5a-8c89-4ba8-cdce-9c67ecc1843e/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1009.290157] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9925e537-e8e4-4ecb-88df-996fa8683609 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.297903] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529ddd5a-8c89-4ba8-cdce-9c67ecc1843e/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1009.298049] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529ddd5a-8c89-4ba8-cdce-9c67ecc1843e/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1009.298262] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b0759553-b575-4cda-93e2-3b4e365d40ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.559471] env[61986]: INFO nova.compute.resource_tracker [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating resource usage from migration a54e6fce-6853-4631-9c9f-715c61d45e67 [ 1009.583419] env[61986]: DEBUG oslo_vmware.rw_handles [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529ddd5a-8c89-4ba8-cdce-9c67ecc1843e/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1009.584828] env[61986]: INFO nova.virt.vmwareapi.images [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Downloaded image file data 24ded09f-c9f0-435c-93fc-1aef721bbca3 [ 1009.589382] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab8a6bf-afc0-4739-a184-642f3195d574 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.610150] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7047d503-0841-466f-8a2a-da6aad15ac77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.640755] env[61986]: INFO nova.virt.vmwareapi.images [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] The imported VM was unregistered [ 1009.646021] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1009.646021] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Creating directory with path [datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.646021] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05b23b89-a230-47e8-a5fb-92d0893195f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.657042] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Created directory with path [datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.657042] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b/OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b.vmdk to [datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk. {{(pid=61986) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1009.657489] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-6d847997-cf44-4664-8995-d2be79ae4e64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.665602] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1009.665602] env[61986]: value = "task-1160261" [ 1009.665602] env[61986]: _type = "Task" [ 1009.665602] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.678810] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160261, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.816076] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e488826a-4b4f-42a2-a8ca-a7eba88b0291 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.825065] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e75f1f-1cd7-46b7-9eb5-5bf18ae77996 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.861261] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee810d2-77b7-4fdb-9ba9-d623aa516a6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.867273] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "18893690-92d4-4391-a395-5cf2115397c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.867557] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "18893690-92d4-4391-a395-5cf2115397c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.875983] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a891bac3-8e5f-4300-bfe8-9448d33407be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.894442] env[61986]: DEBUG nova.compute.provider_tree [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.183451] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160261, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.257220] env[61986]: DEBUG nova.compute.manager [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.257436] env[61986]: DEBUG nova.compute.manager [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing instance network info cache due to event network-changed-74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1010.257767] env[61986]: DEBUG oslo_concurrency.lockutils [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.257957] env[61986]: DEBUG oslo_concurrency.lockutils [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.260054] env[61986]: DEBUG nova.network.neutron [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.370352] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.398587] env[61986]: DEBUG nova.scheduler.client.report [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.685044] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160261, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.725820] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "aa12919e-3891-4da9-a280-4155137864e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.726626] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa12919e-3891-4da9-a280-4155137864e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.727938] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "aa12919e-3891-4da9-a280-4155137864e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.728310] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa12919e-3891-4da9-a280-4155137864e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.731452] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa12919e-3891-4da9-a280-4155137864e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.741540] env[61986]: INFO nova.compute.manager [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Terminating instance [ 1010.742811] env[61986]: DEBUG nova.compute.manager [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.743109] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.744297] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570328f5-a713-4002-8d5d-e93cfd9c78e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.756900] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.761117] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-62f0232a-cecf-4308-993e-716be3e1f2e4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.769154] env[61986]: DEBUG oslo_vmware.api [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 1010.769154] env[61986]: value = "task-1160262" [ 1010.769154] env[61986]: _type = "Task" [ 1010.769154] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.785715] env[61986]: DEBUG oslo_vmware.api [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160262, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.904604] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.369s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.904821] env[61986]: INFO nova.compute.manager [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Migrating [ 1010.913538] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.913865] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.534s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.915846] env[61986]: INFO nova.compute.claims [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.949383] env[61986]: DEBUG nova.network.neutron [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Successfully updated port: 7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.114753] env[61986]: DEBUG nova.network.neutron [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updated VIF entry in instance network info cache for port 74f613a6-57a3-4f00-b018-8c58f254200c. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.115537] env[61986]: DEBUG nova.network.neutron [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.183296] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160261, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.286010] env[61986]: DEBUG oslo_vmware.api [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160262, 'name': PowerOffVM_Task, 'duration_secs': 0.224184} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.286724] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.286994] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1011.287368] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef79f137-9e58-48ae-9f4e-c16d62f68632 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.360657] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.361094] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.361251] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleting the datastore file [datastore1] aa12919e-3891-4da9-a280-4155137864e1 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.361561] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d35221e-f357-44d1-9048-cf8d25767376 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.372497] env[61986]: DEBUG oslo_vmware.api [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for the task: (returnval){ [ 1011.372497] env[61986]: value = "task-1160264" [ 1011.372497] env[61986]: _type = "Task" [ 1011.372497] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.387171] env[61986]: DEBUG oslo_vmware.api [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160264, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.430673] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.430927] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.431247] env[61986]: DEBUG nova.network.neutron [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.453409] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.617892] env[61986]: DEBUG oslo_concurrency.lockutils [req-154f4272-30db-408e-9790-47c50d380c3e req-a34c9475-cfff-4e81-a6f2-55a4ba1233b7 service nova] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.618376] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.618598] env[61986]: DEBUG nova.network.neutron [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.683599] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160261, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.885538] env[61986]: DEBUG oslo_vmware.api [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Task: {'id': task-1160264, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259437} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.885823] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.886021] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.886222] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.886404] env[61986]: INFO nova.compute.manager [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] [instance: aa12919e-3891-4da9-a280-4155137864e1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1011.886661] env[61986]: DEBUG oslo.service.loopingcall [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.886864] env[61986]: DEBUG nova.compute.manager [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.887211] env[61986]: DEBUG nova.network.neutron [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1012.169656] env[61986]: WARNING nova.network.neutron [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] 44bde880-8646-44c4-8c3c-c1faf7d4a9ce already exists in list: networks containing: ['44bde880-8646-44c4-8c3c-c1faf7d4a9ce']. ignoring it [ 1012.183425] env[61986]: DEBUG nova.network.neutron [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance_info_cache with network_info: [{"id": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "address": "fa:16:3e:c3:f2:46", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f71edfb-58", "ovs_interfaceid": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.203561] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160261, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.402081} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.204745] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b/OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b.vmdk to [datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk. [ 1012.205057] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Cleaning up location [datastore2] OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1012.205277] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_17e9510b-128a-4856-8536-b0dbaa263a4b {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.205900] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f63a4f97-0596-4870-8b27-51cbbf5201c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.215225] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1012.215225] env[61986]: value = "task-1160265" [ 1012.215225] env[61986]: _type = "Task" [ 1012.215225] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.231318] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.286309] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb538ab5-8c47-4e83-b7eb-a4ec9cb509d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.295691] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c260c4-11dc-47cf-9c66-e15d1d230573 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.335121] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5a469c-ca56-4c88-8eb6-87efa1c1953e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.341563] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29c4551-29b3-463f-84e3-278b0a3ca9c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.348806] env[61986]: DEBUG nova.compute.manager [req-d363f4ff-f237-4952-ba74-c69ecdfde8f3 req-f8e43d7d-a011-4981-845e-c6a08c5de599 service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Received event network-vif-deleted-038ca95f-90ee-4bd7-afb9-b514378d344e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.349011] env[61986]: INFO nova.compute.manager [req-d363f4ff-f237-4952-ba74-c69ecdfde8f3 req-f8e43d7d-a011-4981-845e-c6a08c5de599 service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Neutron deleted interface 038ca95f-90ee-4bd7-afb9-b514378d344e; detaching it from the instance and deleting it from the info cache [ 1012.350715] env[61986]: DEBUG nova.network.neutron [req-d363f4ff-f237-4952-ba74-c69ecdfde8f3 req-f8e43d7d-a011-4981-845e-c6a08c5de599 service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.362792] env[61986]: DEBUG nova.compute.provider_tree [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.425866] env[61986]: DEBUG nova.compute.manager [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-vif-plugged-7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.426167] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.426394] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] Lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.426663] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] Lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.426879] env[61986]: DEBUG nova.compute.manager [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] No waiting events found dispatching network-vif-plugged-7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1012.427310] env[61986]: WARNING nova.compute.manager [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received unexpected event network-vif-plugged-7228964a-4136-4e76-8fe4-a21ab14deacf for instance with vm_state active and task_state None. [ 1012.427547] env[61986]: DEBUG nova.compute.manager [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-changed-7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.427772] env[61986]: DEBUG nova.compute.manager [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing instance network info cache due to event network-changed-7228964a-4136-4e76-8fe4-a21ab14deacf. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1012.427996] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.575218] env[61986]: DEBUG nova.network.neutron [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7228964a-4136-4e76-8fe4-a21ab14deacf", "address": "fa:16:3e:29:eb:9d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7228964a-41", "ovs_interfaceid": "7228964a-4136-4e76-8fe4-a21ab14deacf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.687844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.728048] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.06607} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.728048] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.728290] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.728488] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk to [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1012.728824] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06f7be5e-40ba-443f-b19f-90754bcbd861 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.734620] env[61986]: DEBUG nova.network.neutron [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.738228] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1012.738228] env[61986]: value = "task-1160266" [ 1012.738228] env[61986]: _type = "Task" [ 1012.738228] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.748861] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160266, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.853041] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b7b9c6d-c9e9-4979-bfb6-f31559a27db0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.864553] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce26a89f-37c4-4407-9fe3-2e244e462196 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.876802] env[61986]: DEBUG nova.scheduler.client.report [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.900231] env[61986]: DEBUG nova.compute.manager [req-d363f4ff-f237-4952-ba74-c69ecdfde8f3 req-f8e43d7d-a011-4981-845e-c6a08c5de599 service nova] [instance: aa12919e-3891-4da9-a280-4155137864e1] Detach interface failed, port_id=038ca95f-90ee-4bd7-afb9-b514378d344e, reason: Instance aa12919e-3891-4da9-a280-4155137864e1 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1013.078267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.079074] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.079248] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.079543] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.079731] env[61986]: DEBUG nova.network.neutron [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Refreshing network info cache for port 7228964a-4136-4e76-8fe4-a21ab14deacf {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1013.081602] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2c977d-ea10-48d3-a73e-2cd43ca0d4ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.101466] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.101819] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.101934] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.102191] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.102523] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.102907] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.103156] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.103333] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.103505] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.103673] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.103879] env[61986]: DEBUG nova.virt.hardware [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.111306] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Reconfiguring VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1013.112156] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bda9db2b-85f2-43f5-9ad2-32394dc2321e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.133496] env[61986]: DEBUG oslo_vmware.api [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 1013.133496] env[61986]: value = "task-1160267" [ 1013.133496] env[61986]: _type = "Task" [ 1013.133496] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.143595] env[61986]: DEBUG oslo_vmware.api [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160267, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.237669] env[61986]: INFO nova.compute.manager [-] [instance: aa12919e-3891-4da9-a280-4155137864e1] Took 1.35 seconds to deallocate network for instance. [ 1013.258875] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160266, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.382379] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.383159] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.387226] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.965s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.387563] env[61986]: DEBUG nova.objects.instance [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'resources' on Instance uuid 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.645406] env[61986]: DEBUG oslo_vmware.api [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.753739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.754087] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160266, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.838510] env[61986]: DEBUG nova.network.neutron [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updated VIF entry in instance network info cache for port 7228964a-4136-4e76-8fe4-a21ab14deacf. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.838970] env[61986]: DEBUG nova.network.neutron [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7228964a-4136-4e76-8fe4-a21ab14deacf", "address": "fa:16:3e:29:eb:9d", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7228964a-41", "ovs_interfaceid": "7228964a-4136-4e76-8fe4-a21ab14deacf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.891610] env[61986]: DEBUG nova.compute.utils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.893854] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.894088] env[61986]: DEBUG nova.network.neutron [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1013.945462] env[61986]: DEBUG nova.policy [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c40c9363dc7741a3bd6a040f20284837', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e0f7d6b6be498eaec797aa64aaea7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1014.151107] env[61986]: DEBUG oslo_vmware.api [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160267, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.155659] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03a7d59-72bf-4b27-81b0-704e7367799d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.166607] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ab0948-e451-44a4-bf40-5811a2a5a884 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.205997] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0874d80c-8a51-41a5-a4f0-2e5a7c53daad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.209983] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0916a61d-d784-476c-958d-22a6fb61b7e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.229638] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1014.237683] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2908f972-c452-4754-a7e0-bcbcc9e4234f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.251010] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160266, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.260034] env[61986]: DEBUG nova.compute.provider_tree [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.266716] env[61986]: DEBUG nova.network.neutron [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Successfully created port: 66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.342847] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff30ed2-95f2-4c33-8483-97aee4ed8c72 req-0aeb54a1-9a36-4981-a2e2-2de2be070c3d service nova] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.400734] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.627193] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524967b0-4fdf-cd40-485b-5e11ee0b1d7c/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1014.628256] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959c9a0a-2e17-45bf-a231-c8da217cd8bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.636692] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524967b0-4fdf-cd40-485b-5e11ee0b1d7c/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1014.636939] env[61986]: ERROR oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524967b0-4fdf-cd40-485b-5e11ee0b1d7c/disk-0.vmdk due to incomplete transfer. [ 1014.640523] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-61974f83-2a8b-466f-b932-b22bfe621fe4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.650184] env[61986]: DEBUG oslo_vmware.api [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160267, 'name': ReconfigVM_Task, 'duration_secs': 1.023458} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.651670] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.651900] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Reconfigured VM to attach interface {{(pid=61986) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1014.654380] env[61986]: DEBUG oslo_vmware.rw_handles [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524967b0-4fdf-cd40-485b-5e11ee0b1d7c/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1014.654579] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Uploaded image 15c86942-389c-4a86-952d-179d0c1d0572 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1014.656974] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1014.657484] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-32d540f6-2f89-4754-b8fa-5cbcd6c05229 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.664054] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1014.664054] env[61986]: value = "task-1160268" [ 1014.664054] env[61986]: _type = "Task" [ 1014.664054] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.672453] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160268, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.739257] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1014.740046] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-037f4725-d39c-4d01-8e00-53a1751210c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.751284] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1014.751284] env[61986]: value = "task-1160269" [ 1014.751284] env[61986]: _type = "Task" [ 1014.751284] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.754519] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160266, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.765274] env[61986]: DEBUG nova.scheduler.client.report [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.768805] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160269, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.159651] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cebb6858-8e17-481f-a2c4-1d08b4ff4d3d tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bda7750f-eec9-40d9-ace5-18d48234126e-7228964a-4136-4e76-8fe4-a21ab14deacf" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.237s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.177752] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160268, 'name': Destroy_Task, 'duration_secs': 0.370629} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.178917] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Destroyed the VM [ 1015.179305] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1015.179695] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6b3d69ba-66e5-4487-9c8e-2df983fd22c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.193482] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1015.193482] env[61986]: value = "task-1160270" [ 1015.193482] env[61986]: _type = "Task" [ 1015.193482] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.927398] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.540s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.930132] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.932586] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160270, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.932772] env[61986]: WARNING oslo_vmware.common.loopingcall [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] task run outlasted interval by 0.239035 sec [ 1015.934730] env[61986]: DEBUG nova.network.neutron [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Successfully updated port: 66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.942092] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.124s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.943593] env[61986]: INFO nova.compute.claims [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.946897] env[61986]: DEBUG nova.compute.manager [req-d6888bfd-005b-4a5d-8923-971130d22e7f req-c454770e-28be-4de4-9e1a-e6435ba745ac service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Received event network-vif-plugged-66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1015.947136] env[61986]: DEBUG oslo_concurrency.lockutils [req-d6888bfd-005b-4a5d-8923-971130d22e7f req-c454770e-28be-4de4-9e1a-e6435ba745ac service nova] Acquiring lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.947352] env[61986]: DEBUG oslo_concurrency.lockutils [req-d6888bfd-005b-4a5d-8923-971130d22e7f req-c454770e-28be-4de4-9e1a-e6435ba745ac service nova] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.947538] env[61986]: DEBUG oslo_concurrency.lockutils [req-d6888bfd-005b-4a5d-8923-971130d22e7f req-c454770e-28be-4de4-9e1a-e6435ba745ac service nova] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.947734] env[61986]: DEBUG nova.compute.manager [req-d6888bfd-005b-4a5d-8923-971130d22e7f req-c454770e-28be-4de4-9e1a-e6435ba745ac service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] No waiting events found dispatching network-vif-plugged-66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.947876] env[61986]: WARNING nova.compute.manager [req-d6888bfd-005b-4a5d-8923-971130d22e7f req-c454770e-28be-4de4-9e1a-e6435ba745ac service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Received unexpected event network-vif-plugged-66980c09-79a0-4d25-aa4e-fd001f6a7c50 for instance with vm_state building and task_state spawning. [ 1015.957527] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160269, 'name': PowerOffVM_Task, 'duration_secs': 0.231337} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.964500] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1015.964711] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1015.969589] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160266, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.543935} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.969772] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160270, 'name': RemoveSnapshot_Task, 'duration_secs': 0.348766} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.972605] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/24ded09f-c9f0-435c-93fc-1aef721bbca3/24ded09f-c9f0-435c-93fc-1aef721bbca3.vmdk to [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1015.972957] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1015.973253] env[61986]: DEBUG nova.compute.manager [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.974465] env[61986]: INFO nova.scheduler.client.report [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted allocations for instance 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f [ 1015.977929] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bb9fe5-6b7f-4531-ad6a-cb693360a3e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.979967] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da18635-c9ce-4e74-8cee-0bd58d8ff36e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.984470] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.984716] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.984871] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.985107] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.985239] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.985605] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.985605] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.985783] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.985891] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.986075] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.986259] env[61986]: DEBUG nova.virt.hardware [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.989898] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d1a1d7-fa30-40c0-8d41-b028748c4efc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.020727] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.021937] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e75b15fc-850d-4668-8472-6a643f09408f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.036905] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999f94e4-a70f-415d-a7c2-83d3d1b2334e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.053407] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1016.053407] env[61986]: value = "task-1160271" [ 1016.053407] env[61986]: _type = "Task" [ 1016.053407] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.060982] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160271, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.452653] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-db9050b8-897e-4d62-8dc2-3b334ab8afd4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.452949] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-db9050b8-897e-4d62-8dc2-3b334ab8afd4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.452949] env[61986]: DEBUG nova.network.neutron [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.474547] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.476020] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.476355] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.476355] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.481845] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dfcf8c7-acf6-4e8c-bbd1-29d03d1ec15f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.500160] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1016.500160] env[61986]: value = "task-1160272" [ 1016.500160] env[61986]: _type = "Task" [ 1016.500160] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.500160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4009024d-54a4-4e67-9511-7f2a8062840f tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.996s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.508917] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160272, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.526503] env[61986]: INFO nova.compute.manager [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Shelve offloading [ 1016.528998] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.528998] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7532ebd-4c4f-40e7-9846-afc6aad9ecaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.536285] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1016.536285] env[61986]: value = "task-1160273" [ 1016.536285] env[61986]: _type = "Task" [ 1016.536285] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.546555] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1016.546792] env[61986]: DEBUG nova.compute.manager [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.548144] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7837b62e-798f-4ad2-9699-43cb43b19b82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.556852] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.556967] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.557162] env[61986]: DEBUG nova.network.neutron [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.565794] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160271, 'name': ReconfigVM_Task, 'duration_secs': 0.282631} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.566575] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Reconfigured VM instance instance-00000036 to attach disk [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19/ee47cab9-5dfd-48ce-ba70-cb800d735b19.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.567188] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80179146-8166-4b1f-91e9-09f3bac9d2d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.573856] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1016.573856] env[61986]: value = "task-1160274" [ 1016.573856] env[61986]: _type = "Task" [ 1016.573856] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.582268] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160274, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.987018] env[61986]: DEBUG nova.network.neutron [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.019047] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160272, 'name': ReconfigVM_Task, 'duration_secs': 0.130914} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.019437] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1017.091702] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160274, 'name': Rename_Task, 'duration_secs': 0.13938} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.092655] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1017.092872] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cae65d7-8660-40fd-a6d1-a090ecbc7a85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.098877] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1017.098877] env[61986]: value = "task-1160275" [ 1017.098877] env[61986]: _type = "Task" [ 1017.098877] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.106953] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.154815] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "interface-bda7750f-eec9-40d9-ace5-18d48234126e-7228964a-4136-4e76-8fe4-a21ab14deacf" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.155180] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bda7750f-eec9-40d9-ace5-18d48234126e-7228964a-4136-4e76-8fe4-a21ab14deacf" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.178644] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e13614f-d071-44cb-b0d6-429e3cdbc65d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.184417] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36321a8d-b254-439b-a2d3-c0afbe4895e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.219457] env[61986]: DEBUG nova.network.neutron [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Updating instance_info_cache with network_info: [{"id": "66980c09-79a0-4d25-aa4e-fd001f6a7c50", "address": "fa:16:3e:3c:40:17", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66980c09-79", "ovs_interfaceid": "66980c09-79a0-4d25-aa4e-fd001f6a7c50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.222546] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb3f58e-8116-40b8-8131-9c6cb55ce34f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.233735] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8736ab8-803d-499f-89e0-e61282afadb6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.247979] env[61986]: DEBUG nova.compute.provider_tree [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.361065] env[61986]: DEBUG nova.network.neutron [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.532416] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.532416] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.532916] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.532916] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.533433] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.533433] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.533671] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.533955] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.534209] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.534404] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.534643] env[61986]: DEBUG nova.virt.hardware [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.541180] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1017.541550] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c6bfcee-ca81-48d9-a9f3-ba529ad13c73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.563271] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1017.563271] env[61986]: value = "task-1160276" [ 1017.563271] env[61986]: _type = "Task" [ 1017.563271] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.572786] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.609091] env[61986]: DEBUG oslo_vmware.api [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160275, 'name': PowerOnVM_Task, 'duration_secs': 0.489707} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.609326] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1017.659012] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.659202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.660701] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa2c240-4f06-4dfd-9f3b-7981aa79d032 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.685846] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64157ff-e23a-403f-8edc-5e617e5409e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.728533] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Reconfiguring VM to detach interface {{(pid=61986) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1017.729746] env[61986]: DEBUG nova.compute.manager [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.730238] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-db9050b8-897e-4d62-8dc2-3b334ab8afd4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.730566] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Instance network_info: |[{"id": "66980c09-79a0-4d25-aa4e-fd001f6a7c50", "address": "fa:16:3e:3c:40:17", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66980c09-79", "ovs_interfaceid": "66980c09-79a0-4d25-aa4e-fd001f6a7c50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.730810] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1aeab9b6-ca42-450e-a3fd-d1d45f880075 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.743809] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b8f9f8-f9d4-4991-b45d-36cd6572c09c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.746636] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:40:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66980c09-79a0-4d25-aa4e-fd001f6a7c50', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.753776] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating folder: Project (b2e0f7d6b6be498eaec797aa64aaea7d). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1017.754965] env[61986]: DEBUG nova.scheduler.client.report [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.757786] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-670cd999-e109-469d-8cc1-f1657f347387 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.766472] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 1017.766472] env[61986]: value = "task-1160278" [ 1017.766472] env[61986]: _type = "Task" [ 1017.766472] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.771271] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created folder: Project (b2e0f7d6b6be498eaec797aa64aaea7d) in parent group-v252271. [ 1017.771456] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating folder: Instances. Parent ref: group-v252438. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1017.772198] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eae11581-53a8-47a6-a1db-4b865638180c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.776657] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.785769] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created folder: Instances in parent group-v252438. [ 1017.785926] env[61986]: DEBUG oslo.service.loopingcall [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.786135] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1017.786772] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff09e52e-cd87-4875-9d00-f77be29f7dc5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.808932] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.808932] env[61986]: value = "task-1160280" [ 1017.808932] env[61986]: _type = "Task" [ 1017.808932] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.819504] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160280, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.856209] env[61986]: DEBUG nova.compute.manager [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Received event network-changed-66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.857103] env[61986]: DEBUG nova.compute.manager [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Refreshing instance network info cache due to event network-changed-66980c09-79a0-4d25-aa4e-fd001f6a7c50. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1017.857103] env[61986]: DEBUG oslo_concurrency.lockutils [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] Acquiring lock "refresh_cache-db9050b8-897e-4d62-8dc2-3b334ab8afd4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.857103] env[61986]: DEBUG oslo_concurrency.lockutils [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] Acquired lock "refresh_cache-db9050b8-897e-4d62-8dc2-3b334ab8afd4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.857103] env[61986]: DEBUG nova.network.neutron [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Refreshing network info cache for port 66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.864553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.913166] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.913432] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.073811] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160276, 'name': ReconfigVM_Task, 'duration_secs': 0.190806} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.074128] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1018.075047] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fc0cb6-4d08-44fe-892c-ecda4cc12e51 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.099969] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07/eb10ea87-8019-4f3b-af3f-32a901ec6a07.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.100324] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bfa79b3-346b-47b7-a7c5-ebc907c73444 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.118017] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1018.118017] env[61986]: value = "task-1160281" [ 1018.118017] env[61986]: _type = "Task" [ 1018.118017] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.126634] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160281, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.234650] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1018.235659] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e842f09-adba-449a-951e-473d2963dd6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.243732] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1018.244016] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5689c37-6765-44e5-9143-f9fd77104bb3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.261609] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.262174] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.264965] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.322s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.265242] env[61986]: DEBUG nova.objects.instance [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lazy-loading 'resources' on Instance uuid 8a594fe5-6fff-48ab-9f7f-474b2a24a486 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.277913] env[61986]: DEBUG oslo_concurrency.lockutils [None req-979bcbe2-7b0e-4a4d-ad2c-ebdbe4c407ed tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.568s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.285112] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.314954] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1018.315093] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1018.315316] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleting the datastore file [datastore1] 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.315979] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd8e1c87-748f-40c1-a87e-4786fffaa7ad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.321262] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160280, 'name': CreateVM_Task, 'duration_secs': 0.363434} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.321761] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.322491] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.322695] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.323043] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.323317] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa8e3358-91ca-4dc0-b277-4a0b75c70a71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.326112] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1018.326112] env[61986]: value = "task-1160283" [ 1018.326112] env[61986]: _type = "Task" [ 1018.326112] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.333485] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1018.333485] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bb22bb-1c08-f44c-b586-5709221a1783" [ 1018.333485] env[61986]: _type = "Task" [ 1018.333485] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.339725] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.344676] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bb22bb-1c08-f44c-b586-5709221a1783, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.415899] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.590655] env[61986]: DEBUG nova.network.neutron [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Updated VIF entry in instance network info cache for port 66980c09-79a0-4d25-aa4e-fd001f6a7c50. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.591277] env[61986]: DEBUG nova.network.neutron [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Updating instance_info_cache with network_info: [{"id": "66980c09-79a0-4d25-aa4e-fd001f6a7c50", "address": "fa:16:3e:3c:40:17", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66980c09-79", "ovs_interfaceid": "66980c09-79a0-4d25-aa4e-fd001f6a7c50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.628419] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160281, 'name': ReconfigVM_Task, 'duration_secs': 0.411069} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.628646] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Reconfigured VM instance instance-00000059 to attach disk [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07/eb10ea87-8019-4f3b-af3f-32a901ec6a07.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.628918] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 50 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1018.770905] env[61986]: DEBUG nova.compute.utils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.772103] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1018.772297] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1018.787497] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.823795] env[61986]: DEBUG nova.policy [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8be836eb63c544e4ab5a831b3397a898', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c64e7f5b4b64d1988edf2a2fcbe32eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1018.840951] env[61986]: DEBUG oslo_vmware.api [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143487} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.841222] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.841417] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.841699] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.853778] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bb22bb-1c08-f44c-b586-5709221a1783, 'name': SearchDatastore_Task, 'duration_secs': 0.018289} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.854113] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.854396] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.854578] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.854749] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.854950] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.858172] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-123f2462-c88c-4e87-a3af-4760b1a426c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.864963] env[61986]: INFO nova.scheduler.client.report [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted allocations for instance 221e10a3-da31-410c-80f8-4bcc2c515710 [ 1018.869906] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.873018] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1018.873018] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efdde063-a31b-4145-88c1-452d111aaf76 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.881351] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1018.881351] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528a566e-77c2-860e-29f7-fb389db0aeef" [ 1018.881351] env[61986]: _type = "Task" [ 1018.881351] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.892232] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528a566e-77c2-860e-29f7-fb389db0aeef, 'name': SearchDatastore_Task, 'duration_secs': 0.009208} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.896480] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfccc0d9-2476-49f2-bc81-fc046448761c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.901587] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1018.901587] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c7906b-b0ed-3b3e-6a15-e925b2dfec85" [ 1018.901587] env[61986]: _type = "Task" [ 1018.901587] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.919607] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c7906b-b0ed-3b3e-6a15-e925b2dfec85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.934731] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.013530] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a5108c-fe16-41a8-9aba-714e1b1fd5f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.021734] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d8122a-f6a6-47d9-b78c-ffa9d151b8d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.052135] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8c4fd0-f84c-4f19-bd54-cb3035a784f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.059436] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d8e616-5f51-467e-b044-1cab82457571 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.072480] env[61986]: DEBUG nova.compute.provider_tree [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.093682] env[61986]: DEBUG oslo_concurrency.lockutils [req-5f4ce25c-aa9f-42d9-900b-5a2045f40d84 req-c2861d0b-86f2-4b4a-a481-7c683f8a5e4a service nova] Releasing lock "refresh_cache-db9050b8-897e-4d62-8dc2-3b334ab8afd4" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.095561] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Successfully created port: c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.136415] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d380def5-6fe6-45c8-a48e-b473f3c817db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.157584] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b452239-34be-441d-8d2b-0f6a326db53f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.175487] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 67 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1019.275665] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.287387] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.324673] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Successfully created port: 25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.371536] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.373176] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1901d057-ed50-4231-ae33-9ee45af3e23e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.380919] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Suspending the VM {{(pid=61986) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1019.381159] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-3e1484c0-c8b7-4ca8-8b1c-90a8bb570b5d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.387575] env[61986]: DEBUG oslo_vmware.api [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1019.387575] env[61986]: value = "task-1160284" [ 1019.387575] env[61986]: _type = "Task" [ 1019.387575] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.396047] env[61986]: DEBUG oslo_vmware.api [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160284, 'name': SuspendVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.413605] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c7906b-b0ed-3b3e-6a15-e925b2dfec85, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.413910] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.414204] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] db9050b8-897e-4d62-8dc2-3b334ab8afd4/db9050b8-897e-4d62-8dc2-3b334ab8afd4.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1019.414500] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8188236-7875-463d-a278-3616baca7998 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.424497] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1019.424497] env[61986]: value = "task-1160285" [ 1019.424497] env[61986]: _type = "Task" [ 1019.424497] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.435782] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160285, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.575955] env[61986]: DEBUG nova.scheduler.client.report [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.725819] env[61986]: DEBUG nova.network.neutron [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Port 9f71edfb-58a5-43cb-9d35-62cda5bf87ad binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1019.802698] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.898984] env[61986]: DEBUG oslo_vmware.api [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160284, 'name': SuspendVM_Task} progress is 58%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.936233] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160285, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.056558] env[61986]: DEBUG nova.compute.manager [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-vif-unplugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.056933] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.057323] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.057587] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.057842] env[61986]: DEBUG nova.compute.manager [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] No waiting events found dispatching network-vif-unplugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.058106] env[61986]: WARNING nova.compute.manager [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received unexpected event network-vif-unplugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 for instance with vm_state shelved_offloaded and task_state unshelving. [ 1020.058375] env[61986]: DEBUG nova.compute.manager [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.058850] env[61986]: DEBUG nova.compute.manager [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing instance network info cache due to event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.059948] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.059948] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.059948] env[61986]: DEBUG nova.network.neutron [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.081793] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.084553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.886s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.086218] env[61986]: INFO nova.compute.claims [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.103799] env[61986]: INFO nova.scheduler.client.report [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleted allocations for instance 8a594fe5-6fff-48ab-9f7f-474b2a24a486 [ 1020.289123] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.295631] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.317011] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.317283] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.317446] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.317640] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.317806] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.317956] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.318185] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.318378] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.318584] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.318757] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.318930] env[61986]: DEBUG nova.virt.hardware [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.319839] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698437ac-0501-47f1-9f8a-04daaddeeecd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.327516] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf036f35-6939-4b7a-a577-799696ae28a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.397963] env[61986]: DEBUG oslo_vmware.api [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160284, 'name': SuspendVM_Task, 'duration_secs': 0.701303} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.398110] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Suspended the VM {{(pid=61986) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1020.398245] env[61986]: DEBUG nova.compute.manager [None req-8e47efce-eeb3-4b93-88e4-d0593135bb24 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.399289] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806325a9-2841-4364-ae59-6ebef6e5f794 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.434803] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160285, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709467} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.435167] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] db9050b8-897e-4d62-8dc2-3b334ab8afd4/db9050b8-897e-4d62-8dc2-3b334ab8afd4.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1020.435430] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.435730] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b687e21e-4529-4814-b537-2fc77edbac16 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.442484] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1020.442484] env[61986]: value = "task-1160286" [ 1020.442484] env[61986]: _type = "Task" [ 1020.442484] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.450458] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.614017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-543b123b-8096-4659-b78a-c3cb5c246914 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "8a594fe5-6fff-48ab-9f7f-474b2a24a486" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.467s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.752706] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.752984] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.753161] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.788772] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.888652] env[61986]: DEBUG nova.network.neutron [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updated VIF entry in instance network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.889232] env[61986]: DEBUG nova.network.neutron [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.953756] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065807} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.955435] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.959228] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caaa51b0-5967-46d1-94bb-90ce2da6b802 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.964909] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.987413] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] db9050b8-897e-4d62-8dc2-3b334ab8afd4/db9050b8-897e-4d62-8dc2-3b334ab8afd4.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.988965] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Successfully updated port: c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.993198] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfaeb696-3a4e-4fef-8fb0-5299c658378a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.016103] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1021.016103] env[61986]: value = "task-1160287" [ 1021.016103] env[61986]: _type = "Task" [ 1021.016103] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.025500] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160287, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.291851] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.296294] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6878514-366a-45b4-a44b-f85c3e130ff0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.303293] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd853a9-b22e-4db4-83ee-950109c16c21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.335177] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7d37ff-bddf-460c-9de6-9fd831174ece {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.343074] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bb241a-92aa-4d94-87e4-cc834dd88b19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.356442] env[61986]: DEBUG nova.compute.provider_tree [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.394855] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce72154d-5b6d-4ed1-ab64-f5a269b8b13c req-05b7d6c5-d241-4a85-9e77-2ed3f3fb8786 service nova] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.464042] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "45d80bab-0140-4473-9dc2-4a5481838aed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.464164] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "45d80bab-0140-4473-9dc2-4a5481838aed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.464386] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "45d80bab-0140-4473-9dc2-4a5481838aed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.464576] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "45d80bab-0140-4473-9dc2-4a5481838aed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.464802] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "45d80bab-0140-4473-9dc2-4a5481838aed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.466997] env[61986]: INFO nova.compute.manager [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Terminating instance [ 1021.468781] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "refresh_cache-45d80bab-0140-4473-9dc2-4a5481838aed" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.468973] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquired lock "refresh_cache-45d80bab-0140-4473-9dc2-4a5481838aed" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.469173] env[61986]: DEBUG nova.network.neutron [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.526681] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160287, 'name': ReconfigVM_Task, 'duration_secs': 0.266816} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.526813] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Reconfigured VM instance instance-0000005a to attach disk [datastore2] db9050b8-897e-4d62-8dc2-3b334ab8afd4/db9050b8-897e-4d62-8dc2-3b334ab8afd4.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.527446] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf61d379-886e-4fdd-a3ab-06c6b93d8379 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.534560] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1021.534560] env[61986]: value = "task-1160288" [ 1021.534560] env[61986]: _type = "Task" [ 1021.534560] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.546968] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160288, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.789861] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.790800] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.790959] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.791151] env[61986]: DEBUG nova.network.neutron [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.858996] env[61986]: DEBUG nova.scheduler.client.report [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.986383] env[61986]: DEBUG nova.network.neutron [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1021.991616] env[61986]: INFO nova.compute.manager [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Resuming [ 1021.992202] env[61986]: DEBUG nova.objects.instance [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lazy-loading 'flavor' on Instance uuid ee47cab9-5dfd-48ce-ba70-cb800d735b19 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.031412] env[61986]: DEBUG nova.network.neutron [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.046798] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160288, 'name': Rename_Task, 'duration_secs': 0.150908} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.047075] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1022.047318] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d11f5ed2-e28c-4723-bb7b-57b1a96eecd8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.054042] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1022.054042] env[61986]: value = "task-1160289" [ 1022.054042] env[61986]: _type = "Task" [ 1022.054042] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.062990] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160289, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.102922] env[61986]: DEBUG nova.compute.manager [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received event network-vif-plugged-c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.103110] env[61986]: DEBUG oslo_concurrency.lockutils [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] Acquiring lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.103325] env[61986]: DEBUG oslo_concurrency.lockutils [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.103399] env[61986]: DEBUG oslo_concurrency.lockutils [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.103536] env[61986]: DEBUG nova.compute.manager [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] No waiting events found dispatching network-vif-plugged-c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.103703] env[61986]: WARNING nova.compute.manager [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received unexpected event network-vif-plugged-c005e588-3e50-435a-849a-95a38ccac917 for instance with vm_state building and task_state spawning. [ 1022.103862] env[61986]: DEBUG nova.compute.manager [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received event network-changed-c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.104024] env[61986]: DEBUG nova.compute.manager [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Refreshing instance network info cache due to event network-changed-c005e588-3e50-435a-849a-95a38ccac917. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.104214] env[61986]: DEBUG oslo_concurrency.lockutils [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] Acquiring lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.104347] env[61986]: DEBUG oslo_concurrency.lockutils [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] Acquired lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.104529] env[61986]: DEBUG nova.network.neutron [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Refreshing network info cache for port c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1022.291099] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.363896] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.364601] env[61986]: DEBUG nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1022.367310] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.768s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.368835] env[61986]: INFO nova.compute.claims [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.534573] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Releasing lock "refresh_cache-45d80bab-0140-4473-9dc2-4a5481838aed" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.534995] env[61986]: DEBUG nova.compute.manager [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.535204] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.536066] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695c8c01-c809-4df0-b93e-10a881ca1a33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.543892] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.544135] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75a67235-9d06-4423-80e7-ee8dee3fc121 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.558634] env[61986]: DEBUG oslo_vmware.api [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 1022.558634] env[61986]: value = "task-1160290" [ 1022.558634] env[61986]: _type = "Task" [ 1022.558634] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.564749] env[61986]: DEBUG oslo_vmware.api [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160289, 'name': PowerOnVM_Task, 'duration_secs': 0.419705} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.565324] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1022.565539] env[61986]: INFO nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1022.565730] env[61986]: DEBUG nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.566804] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a07fbf-80f2-4da8-9d03-896698b93938 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.572624] env[61986]: DEBUG oslo_vmware.api [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.643818] env[61986]: DEBUG nova.network.neutron [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance_info_cache with network_info: [{"id": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "address": "fa:16:3e:c3:f2:46", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f71edfb-58", "ovs_interfaceid": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.660813] env[61986]: DEBUG nova.network.neutron [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.688501] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Successfully updated port: 25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.790251] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.803564] env[61986]: DEBUG nova.network.neutron [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.874283] env[61986]: DEBUG nova.compute.utils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.877530] env[61986]: DEBUG nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1023.001325] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.001559] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquired lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.001771] env[61986]: DEBUG nova.network.neutron [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.068909] env[61986]: DEBUG oslo_vmware.api [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160290, 'name': PowerOffVM_Task, 'duration_secs': 0.120338} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.069201] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.069373] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1023.069880] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07cb8c99-c380-422a-88c0-ec3150cd4b4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.091115] env[61986]: INFO nova.compute.manager [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Took 20.73 seconds to build instance. [ 1023.096072] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1023.096284] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1023.096468] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleting the datastore file [datastore1] 45d80bab-0140-4473-9dc2-4a5481838aed {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.096717] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02dc9352-cb8d-4468-90ef-15b034203458 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.104996] env[61986]: DEBUG oslo_vmware.api [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for the task: (returnval){ [ 1023.104996] env[61986]: value = "task-1160292" [ 1023.104996] env[61986]: _type = "Task" [ 1023.104996] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.113661] env[61986]: DEBUG oslo_vmware.api [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.146207] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.192104] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.291126] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.309062] env[61986]: DEBUG oslo_concurrency.lockutils [req-0aee279b-60db-4c13-98ec-c344c1daade8 req-65ef571b-51c1-44ec-90bd-e5245089dbba service nova] Releasing lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.309432] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.309588] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.381405] env[61986]: DEBUG nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.537526] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.576462] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f60717-e4bf-454a-a200-8284811461ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.584423] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfd3e49-3e99-47f2-8819-30bba6874f7a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.615279] env[61986]: DEBUG oslo_concurrency.lockutils [None req-049ef2c3-e815-4193-a21d-2e3f642d20fb tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.266s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.618102] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.081s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.618328] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.618534] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.618701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.623319] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d019f0bf-3881-473a-b601-f8d34d4be89f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.626040] env[61986]: INFO nova.compute.manager [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Terminating instance [ 1023.628380] env[61986]: DEBUG nova.compute.manager [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.628608] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1023.630164] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a34bf5f-781a-4750-be5e-63b1ef07d3ca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.637902] env[61986]: DEBUG oslo_vmware.api [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Task: {'id': task-1160292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089217} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.639520] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c192689-d153-4d5a-b7c5-29b770474b7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.643629] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.643862] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.644068] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.644260] env[61986]: INFO nova.compute.manager [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1023.644505] env[61986]: DEBUG oslo.service.loopingcall [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.648616] env[61986]: DEBUG nova.compute.manager [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.648743] env[61986]: DEBUG nova.network.neutron [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.650173] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.654097] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31a9b73e-00ff-4fee-ba30-4bea827b3373 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.665303] env[61986]: DEBUG nova.compute.provider_tree [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.671731] env[61986]: DEBUG oslo_vmware.api [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1023.671731] env[61986]: value = "task-1160293" [ 1023.671731] env[61986]: _type = "Task" [ 1023.671731] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.679616] env[61986]: DEBUG oslo_vmware.api [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160293, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.683846] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1276b344-db36-423d-85e4-30f51b9d78a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.686848] env[61986]: DEBUG nova.network.neutron [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.707868] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19ddc6f-178d-4c26-98a7-cc7fb9549c27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.715925] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 83 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1023.791945] env[61986]: DEBUG oslo_vmware.api [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160278, 'name': ReconfigVM_Task, 'duration_secs': 5.751244} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.792645] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.792914] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Reconfigured VM to detach interface {{(pid=61986) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1023.802682] env[61986]: DEBUG nova.network.neutron [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [{"id": "78729401-474d-4e4f-8159-aff2e908be0b", "address": "fa:16:3e:61:fd:5c", "network": {"id": "fa2957cb-4565-4ae4-a271-d4994982414d", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1312500491-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c94c168de68c4608b772cf6fa6012dcb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78729401-47", "ovs_interfaceid": "78729401-474d-4e4f-8159-aff2e908be0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.850754] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1024.130685] env[61986]: DEBUG nova.compute.manager [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received event network-vif-plugged-25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1024.130985] env[61986]: DEBUG oslo_concurrency.lockutils [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] Acquiring lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.131179] env[61986]: DEBUG oslo_concurrency.lockutils [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.131308] env[61986]: DEBUG oslo_concurrency.lockutils [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.131490] env[61986]: DEBUG nova.compute.manager [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] No waiting events found dispatching network-vif-plugged-25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1024.131692] env[61986]: WARNING nova.compute.manager [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received unexpected event network-vif-plugged-25671e16-307f-48e8-b99f-891dea3c0a3a for instance with vm_state building and task_state spawning. [ 1024.131859] env[61986]: DEBUG nova.compute.manager [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received event network-changed-25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1024.132028] env[61986]: DEBUG nova.compute.manager [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Refreshing instance network info cache due to event network-changed-25671e16-307f-48e8-b99f-891dea3c0a3a. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1024.132205] env[61986]: DEBUG oslo_concurrency.lockutils [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] Acquiring lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.168794] env[61986]: DEBUG nova.scheduler.client.report [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.182668] env[61986]: DEBUG oslo_vmware.api [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160293, 'name': PowerOffVM_Task, 'duration_secs': 0.2289} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.182969] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1024.183144] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1024.183487] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a00c44f-8b5c-49e7-9204-da3bfc2ce4d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.190421] env[61986]: DEBUG nova.network.neutron [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.224467] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.224810] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87c3314c-4bcb-40d4-9d04-6f7265588e52 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.232550] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1024.232550] env[61986]: value = "task-1160295" [ 1024.232550] env[61986]: _type = "Task" [ 1024.232550] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.241701] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160295, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.266773] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1024.267072] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1024.267243] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleting the datastore file [datastore2] db9050b8-897e-4d62-8dc2-3b334ab8afd4 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.267684] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-295405c9-4541-4719-adf9-73c34dd38c6d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.279602] env[61986]: DEBUG oslo_vmware.api [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1024.279602] env[61986]: value = "task-1160296" [ 1024.279602] env[61986]: _type = "Task" [ 1024.279602] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.289345] env[61986]: DEBUG oslo_vmware.api [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.305430] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Releasing lock "refresh_cache-ee47cab9-5dfd-48ce-ba70-cb800d735b19" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.306819] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a5bb4d-af29-474b-a1b1-6d7f16c3b42f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.313691] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Resuming the VM {{(pid=61986) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1024.313970] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9d1aed1-3dd7-4490-8f74-d9dbc23ecb70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.320914] env[61986]: DEBUG oslo_vmware.api [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1024.320914] env[61986]: value = "task-1160297" [ 1024.320914] env[61986]: _type = "Task" [ 1024.320914] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.334803] env[61986]: DEBUG oslo_vmware.api [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.354252] env[61986]: DEBUG nova.network.neutron [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Updating instance_info_cache with network_info: [{"id": "c005e588-3e50-435a-849a-95a38ccac917", "address": "fa:16:3e:53:fa:ea", "network": {"id": "a2aa2876-c3fa-4902-ba57-50f984976a70", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1745244369", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc005e588-3e", "ovs_interfaceid": "c005e588-3e50-435a-849a-95a38ccac917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "25671e16-307f-48e8-b99f-891dea3c0a3a", "address": "fa:16:3e:4a:91:91", "network": {"id": "d106cfec-f509-4e54-8d6e-56593a100413", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-340664290", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25671e16-30", "ovs_interfaceid": "25671e16-307f-48e8-b99f-891dea3c0a3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.392205] env[61986]: DEBUG nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1024.422903] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.423387] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.423616] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.423873] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.424198] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.424460] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.424846] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.425082] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.425311] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.425535] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.425773] env[61986]: DEBUG nova.virt.hardware [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.427471] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7e36df-0d0b-47b4-8d1d-b4b8a2b4f152 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.437401] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36c6667-5b62-418e-8815-003f56b88c5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.452318] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.458049] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Creating folder: Project (a408b25b63214968a895a509b7fd9a83). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1024.458301] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e18e2f7c-3fce-4a09-97ae-037084f59f7f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.468245] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Created folder: Project (a408b25b63214968a895a509b7fd9a83) in parent group-v252271. [ 1024.468517] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Creating folder: Instances. Parent ref: group-v252441. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1024.468747] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a85d9d35-fbaf-478d-aa01-c957ace76904 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.479246] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Created folder: Instances in parent group-v252441. [ 1024.479529] env[61986]: DEBUG oslo.service.loopingcall [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.479745] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.479980] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4a68231-c561-42b4-a0cc-8ba144c9bb1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.497482] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.497482] env[61986]: value = "task-1160300" [ 1024.497482] env[61986]: _type = "Task" [ 1024.497482] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.505879] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160300, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.677340] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.677951] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1024.680942] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.768s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.682411] env[61986]: INFO nova.compute.claims [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.692937] env[61986]: INFO nova.compute.manager [-] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Took 1.04 seconds to deallocate network for instance. [ 1024.744112] env[61986]: DEBUG oslo_vmware.api [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160295, 'name': PowerOnVM_Task, 'duration_secs': 0.404644} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.744389] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.744581] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d49e41-68e3-4fa9-b774-584ded60a519 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance 'eb10ea87-8019-4f3b-af3f-32a901ec6a07' progress to 100 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1024.789603] env[61986]: DEBUG oslo_vmware.api [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195728} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.789861] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.790056] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1024.790242] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1024.790417] env[61986]: INFO nova.compute.manager [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1024.790681] env[61986]: DEBUG oslo.service.loopingcall [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.790899] env[61986]: DEBUG nova.compute.manager [-] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1024.790998] env[61986]: DEBUG nova.network.neutron [-] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1024.832856] env[61986]: DEBUG oslo_vmware.api [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160297, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.856595] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Releasing lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.856982] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Instance network_info: |[{"id": "c005e588-3e50-435a-849a-95a38ccac917", "address": "fa:16:3e:53:fa:ea", "network": {"id": "a2aa2876-c3fa-4902-ba57-50f984976a70", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1745244369", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc005e588-3e", "ovs_interfaceid": "c005e588-3e50-435a-849a-95a38ccac917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "25671e16-307f-48e8-b99f-891dea3c0a3a", "address": "fa:16:3e:4a:91:91", "network": {"id": "d106cfec-f509-4e54-8d6e-56593a100413", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-340664290", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25671e16-30", "ovs_interfaceid": "25671e16-307f-48e8-b99f-891dea3c0a3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1024.857320] env[61986]: DEBUG oslo_concurrency.lockutils [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] Acquired lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.857504] env[61986]: DEBUG nova.network.neutron [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Refreshing network info cache for port 25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1024.858777] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:fa:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8459aaf-d6a8-46fb-ad14-464ac3104695', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c005e588-3e50-435a-849a-95a38ccac917', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:91:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25671e16-307f-48e8-b99f-891dea3c0a3a', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.872215] env[61986]: DEBUG oslo.service.loopingcall [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.873378] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.873618] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9150620c-b08d-4929-8eb3-76be5a356afa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.895627] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.895627] env[61986]: value = "task-1160301" [ 1024.895627] env[61986]: _type = "Task" [ 1024.895627] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.906194] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160301, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.009910] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160300, 'name': CreateVM_Task, 'duration_secs': 0.456585} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.011211] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1025.011211] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.011346] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.011626] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.011901] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-770c4985-219a-4004-a55b-f6214666f101 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.016928] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1025.016928] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52170dda-f537-1a35-d54b-097b4d9290e0" [ 1025.016928] env[61986]: _type = "Task" [ 1025.016928] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.024892] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52170dda-f537-1a35-d54b-097b4d9290e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.087329] env[61986]: DEBUG nova.compute.manager [req-e887e40e-758e-4f2e-a4cb-e4c7ad8a8625 req-030a44cc-da39-4d99-b0c9-259d5eae0ea8 service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Received event network-vif-deleted-66980c09-79a0-4d25-aa4e-fd001f6a7c50 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.087541] env[61986]: INFO nova.compute.manager [req-e887e40e-758e-4f2e-a4cb-e4c7ad8a8625 req-030a44cc-da39-4d99-b0c9-259d5eae0ea8 service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Neutron deleted interface 66980c09-79a0-4d25-aa4e-fd001f6a7c50; detaching it from the instance and deleting it from the info cache [ 1025.087709] env[61986]: DEBUG nova.network.neutron [req-e887e40e-758e-4f2e-a4cb-e4c7ad8a8625 req-030a44cc-da39-4d99-b0c9-259d5eae0ea8 service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.091747] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.091949] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquired lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.092137] env[61986]: DEBUG nova.network.neutron [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.188910] env[61986]: DEBUG nova.compute.utils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.193276] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1025.193459] env[61986]: DEBUG nova.network.neutron [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1025.204221] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.279162] env[61986]: DEBUG nova.policy [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8562656b1e46628059ea24f9e16b26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0db18c866dc4ffb8bcb050f8ec6021d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1025.336248] env[61986]: DEBUG oslo_vmware.api [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160297, 'name': PowerOnVM_Task, 'duration_secs': 0.914312} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.336406] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Resumed the VM {{(pid=61986) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1025.336587] env[61986]: DEBUG nova.compute.manager [None req-0e67516a-8c1d-4c46-b603-e9cffab667c9 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.337408] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbff380b-a3e9-4be3-a979-aeb987c51730 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.412286] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160301, 'name': CreateVM_Task, 'duration_secs': 0.389957} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.412542] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1025.413746] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.534269] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52170dda-f537-1a35-d54b-097b4d9290e0, 'name': SearchDatastore_Task, 'duration_secs': 0.010732} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.534613] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.535542] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.535853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.536974] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.536974] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.536974] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.536974] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.537193] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3112dc0-aa64-42a8-ba1e-2eb93770019e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.539811] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fae6485-a06e-4ac3-9ca1-884debc7bb8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.545443] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1025.545443] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d7b7b4-2667-8514-296a-251a7ff23a7d" [ 1025.545443] env[61986]: _type = "Task" [ 1025.545443] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.553521] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.553712] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1025.554784] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa7760cc-7e8e-4118-90be-423649e6f722 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.560422] env[61986]: DEBUG nova.network.neutron [-] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.561600] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d7b7b4-2667-8514-296a-251a7ff23a7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.567913] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1025.567913] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52230946-0a89-81ea-4333-c75aae48e5a8" [ 1025.567913] env[61986]: _type = "Task" [ 1025.567913] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.572309] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52230946-0a89-81ea-4333-c75aae48e5a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.590999] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-971f1a0d-83d0-47ec-ad3f-eb4763d122e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.601579] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b292c9-2775-4073-a110-07a9e4223d6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.628012] env[61986]: DEBUG nova.compute.manager [req-e887e40e-758e-4f2e-a4cb-e4c7ad8a8625 req-030a44cc-da39-4d99-b0c9-259d5eae0ea8 service nova] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Detach interface failed, port_id=66980c09-79a0-4d25-aa4e-fd001f6a7c50, reason: Instance db9050b8-897e-4d62-8dc2-3b334ab8afd4 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1025.693610] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1025.710880] env[61986]: DEBUG nova.network.neutron [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Updated VIF entry in instance network info cache for port 25671e16-307f-48e8-b99f-891dea3c0a3a. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.711620] env[61986]: DEBUG nova.network.neutron [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Updating instance_info_cache with network_info: [{"id": "c005e588-3e50-435a-849a-95a38ccac917", "address": "fa:16:3e:53:fa:ea", "network": {"id": "a2aa2876-c3fa-4902-ba57-50f984976a70", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1745244369", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc005e588-3e", "ovs_interfaceid": "c005e588-3e50-435a-849a-95a38ccac917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "25671e16-307f-48e8-b99f-891dea3c0a3a", "address": "fa:16:3e:4a:91:91", "network": {"id": "d106cfec-f509-4e54-8d6e-56593a100413", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-340664290", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25671e16-30", "ovs_interfaceid": "25671e16-307f-48e8-b99f-891dea3c0a3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.732202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.733314] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bda7750f-eec9-40d9-ace5-18d48234126e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.733314] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.733314] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.733314] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bda7750f-eec9-40d9-ace5-18d48234126e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.736463] env[61986]: INFO nova.compute.manager [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Terminating instance [ 1025.739163] env[61986]: DEBUG nova.compute.manager [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.739679] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1025.740777] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28c0bb4-a724-4795-9988-1105cb502c05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.753722] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1025.753981] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db80b4d3-1a40-4dbf-86b1-6b2ffb0edf08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.776952] env[61986]: DEBUG oslo_vmware.api [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 1025.776952] env[61986]: value = "task-1160302" [ 1025.776952] env[61986]: _type = "Task" [ 1025.776952] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.783848] env[61986]: DEBUG oslo_vmware.api [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160302, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.918336] env[61986]: INFO nova.network.neutron [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Port 7228964a-4136-4e76-8fe4-a21ab14deacf from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1025.918706] env[61986]: DEBUG nova.network.neutron [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [{"id": "74f613a6-57a3-4f00-b018-8c58f254200c", "address": "fa:16:3e:4f:c9:c8", "network": {"id": "44bde880-8646-44c4-8c3c-c1faf7d4a9ce", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-486682452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "617d3b641cdc4d1b815b5edf7e970d62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f613a6-57", "ovs_interfaceid": "74f613a6-57a3-4f00-b018-8c58f254200c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.936149] env[61986]: DEBUG nova.network.neutron [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Successfully created port: b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.968845] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed83a4ff-14da-44bc-8622-ec206b6ca513 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.976468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71be456c-f9e1-4bb7-b1a5-146f127db5ff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.008218] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af807d21-528c-484c-bbed-ad1317d5b964 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.016712] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62ad4b1-5a92-4bbe-bc59-e4f82ea046b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.029772] env[61986]: DEBUG nova.compute.provider_tree [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.055527] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d7b7b4-2667-8514-296a-251a7ff23a7d, 'name': SearchDatastore_Task, 'duration_secs': 0.04114} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.055852] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.059015] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.059015] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.063083] env[61986]: INFO nova.compute.manager [-] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Took 1.27 seconds to deallocate network for instance. [ 1026.077026] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52230946-0a89-81ea-4333-c75aae48e5a8, 'name': SearchDatastore_Task, 'duration_secs': 0.023053} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.078479] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4990e72-419e-488d-aa53-5baf65154fbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.084043] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1026.084043] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e163db-94fd-5ab3-276d-7c0368c80fda" [ 1026.084043] env[61986]: _type = "Task" [ 1026.084043] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.096061] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e163db-94fd-5ab3-276d-7c0368c80fda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.214041] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.214319] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1026.221023] env[61986]: DEBUG oslo_concurrency.lockutils [req-767808ad-a01b-4b57-9a8b-0568902e52c0 req-25a8da1d-94f3-42b8-ab6a-3edcafd054ef service nova] Releasing lock "refresh_cache-645ecfee-3c0b-47c1-86c1-8ca21781438a" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.285514] env[61986]: DEBUG oslo_vmware.api [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160302, 'name': PowerOffVM_Task, 'duration_secs': 0.195948} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.285787] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1026.285961] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1026.286702] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b66b997-1e84-4552-9927-a238ee8b4ced {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.349755] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1026.349940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1026.350142] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleting the datastore file [datastore1] bda7750f-eec9-40d9-ace5-18d48234126e {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.350412] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff2a6dee-fcd5-4d40-a62c-94c52233d7bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.357984] env[61986]: DEBUG oslo_vmware.api [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 1026.357984] env[61986]: value = "task-1160304" [ 1026.357984] env[61986]: _type = "Task" [ 1026.357984] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.365838] env[61986]: DEBUG oslo_vmware.api [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.421372] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Releasing lock "refresh_cache-bda7750f-eec9-40d9-ace5-18d48234126e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.533542] env[61986]: DEBUG nova.scheduler.client.report [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.571900] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.595778] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e163db-94fd-5ab3-276d-7c0368c80fda, 'name': SearchDatastore_Task, 'duration_secs': 0.009354} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.596270] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.599091] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 68626872-0875-4ff1-81ad-85bdff30ac94/68626872-0875-4ff1-81ad-85bdff30ac94.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1026.599091] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.599091] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.599091] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5bcb8854-1086-4f65-afb6-93050409695d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.599578] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75ac1223-2987-4115-85b6-966e3e25d5da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.609019] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1026.609019] env[61986]: value = "task-1160305" [ 1026.609019] env[61986]: _type = "Task" [ 1026.609019] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.611321] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.611797] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1026.613462] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b7ff5b7-780b-4360-81c1-e22e29c34caa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.620054] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.622998] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1026.622998] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5266869e-83da-0bbe-8e32-864c397b95d3" [ 1026.622998] env[61986]: _type = "Task" [ 1026.622998] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.629720] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5266869e-83da-0bbe-8e32-864c397b95d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.703869] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1026.737588] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.737815] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.737995] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.738192] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.738417] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.739152] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.739152] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.739152] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.739345] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.739438] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.739605] env[61986]: DEBUG nova.virt.hardware [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.741062] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69213a42-fce5-4e6a-8230-20851c3cf17a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.749167] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffe5ca9-f066-4258-8dab-aeeebced3003 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.871104] env[61986]: DEBUG oslo_vmware.api [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149972} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.871390] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.871600] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1026.871783] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1026.871963] env[61986]: INFO nova.compute.manager [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1026.872235] env[61986]: DEBUG oslo.service.loopingcall [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.872438] env[61986]: DEBUG nova.compute.manager [-] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.872534] env[61986]: DEBUG nova.network.neutron [-] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1026.926752] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0aac9094-a02a-4398-a750-c2b3d8eb4ed6 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "interface-bda7750f-eec9-40d9-ace5-18d48234126e-7228964a-4136-4e76-8fe4-a21ab14deacf" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.771s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.037925] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.038611] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1027.042213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.289s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.042500] env[61986]: DEBUG nova.objects.instance [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lazy-loading 'resources' on Instance uuid aa12919e-3891-4da9-a280-4155137864e1 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.118415] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160305, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479602} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.118629] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 68626872-0875-4ff1-81ad-85bdff30ac94/68626872-0875-4ff1-81ad-85bdff30ac94.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1027.118844] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.119115] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5679c6f1-4425-47fb-8a5b-1c973a7b8f75 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.127742] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1027.127742] env[61986]: value = "task-1160306" [ 1027.127742] env[61986]: _type = "Task" [ 1027.127742] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.136148] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5266869e-83da-0bbe-8e32-864c397b95d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008555} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.136148] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-729d5249-d3f8-4759-9f12-3af5dc3c3669 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.141308] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160306, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.144206] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1027.144206] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523374b0-e48e-b8da-843a-5aa3b3b28d8b" [ 1027.144206] env[61986]: _type = "Task" [ 1027.144206] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.151666] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523374b0-e48e-b8da-843a-5aa3b3b28d8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.213699] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1027.546645] env[61986]: DEBUG nova.compute.utils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.550976] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1027.551168] env[61986]: DEBUG nova.network.neutron [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1027.614701] env[61986]: DEBUG nova.policy [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b3bfc2d868d403ba0896447954d4305', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86fcc2bbe80645b5a5cde8b7837b7575', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1027.639853] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160306, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06558} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.640171] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.640997] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ef8e20-6a6e-4834-9ef2-7766b1ca1433 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.665860] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 68626872-0875-4ff1-81ad-85bdff30ac94/68626872-0875-4ff1-81ad-85bdff30ac94.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.667210] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36a2923e-36a7-4e70-8a43-2f18db75f532 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.682697] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.682943] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.683156] env[61986]: DEBUG nova.compute.manager [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Going to confirm migration 4 {{(pid=61986) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1027.692280] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523374b0-e48e-b8da-843a-5aa3b3b28d8b, 'name': SearchDatastore_Task, 'duration_secs': 0.010075} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.693014] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.693649] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 645ecfee-3c0b-47c1-86c1-8ca21781438a/645ecfee-3c0b-47c1-86c1-8ca21781438a.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1027.693649] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4456c450-f946-4692-aee8-5d91e02314dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.701128] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1027.701128] env[61986]: value = "task-1160307" [ 1027.701128] env[61986]: _type = "Task" [ 1027.701128] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.705497] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1027.705497] env[61986]: value = "task-1160308" [ 1027.705497] env[61986]: _type = "Task" [ 1027.705497] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.712978] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160307, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.717299] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.721249] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.772509] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eb9d44-3894-4c8f-8159-f0ff15a218d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.783471] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e55882-0b2a-4afd-8a7f-c565026fcca6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.817641] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbcdded-521b-41f8-abce-0319a0d2da51 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.825613] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac4b7d8-5330-46fa-bcd5-09b8bd822ef5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.842160] env[61986]: DEBUG nova.compute.provider_tree [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.844383] env[61986]: DEBUG nova.compute.manager [req-1011db99-1aeb-440d-9be4-1b91216d697b req-35ff4fcc-491a-48a4-b692-9670a6bc227e service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Received event network-vif-plugged-b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.844603] env[61986]: DEBUG oslo_concurrency.lockutils [req-1011db99-1aeb-440d-9be4-1b91216d697b req-35ff4fcc-491a-48a4-b692-9670a6bc227e service nova] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.844857] env[61986]: DEBUG oslo_concurrency.lockutils [req-1011db99-1aeb-440d-9be4-1b91216d697b req-35ff4fcc-491a-48a4-b692-9670a6bc227e service nova] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.845043] env[61986]: DEBUG oslo_concurrency.lockutils [req-1011db99-1aeb-440d-9be4-1b91216d697b req-35ff4fcc-491a-48a4-b692-9670a6bc227e service nova] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.845240] env[61986]: DEBUG nova.compute.manager [req-1011db99-1aeb-440d-9be4-1b91216d697b req-35ff4fcc-491a-48a4-b692-9670a6bc227e service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] No waiting events found dispatching network-vif-plugged-b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.845408] env[61986]: WARNING nova.compute.manager [req-1011db99-1aeb-440d-9be4-1b91216d697b req-35ff4fcc-491a-48a4-b692-9670a6bc227e service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Received unexpected event network-vif-plugged-b4336d9c-969c-4d2b-830f-c988c73c8410 for instance with vm_state building and task_state spawning. [ 1028.007349] env[61986]: DEBUG nova.network.neutron [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Successfully created port: 85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.051441] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1028.077360] env[61986]: DEBUG nova.compute.manager [req-c9eaa1b6-d280-4214-829c-20e08a6346f1 req-4e6217ce-fb9e-42eb-aa18-623ec5c470f0 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Received event network-vif-deleted-74f613a6-57a3-4f00-b018-8c58f254200c {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1028.077455] env[61986]: INFO nova.compute.manager [req-c9eaa1b6-d280-4214-829c-20e08a6346f1 req-4e6217ce-fb9e-42eb-aa18-623ec5c470f0 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Neutron deleted interface 74f613a6-57a3-4f00-b018-8c58f254200c; detaching it from the instance and deleting it from the info cache [ 1028.077608] env[61986]: DEBUG nova.network.neutron [req-c9eaa1b6-d280-4214-829c-20e08a6346f1 req-4e6217ce-fb9e-42eb-aa18-623ec5c470f0 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.215344] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160307, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.222033] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160308, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.289446] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.289654] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.289840] env[61986]: DEBUG nova.network.neutron [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.290048] env[61986]: DEBUG nova.objects.instance [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lazy-loading 'info_cache' on Instance uuid eb10ea87-8019-4f3b-af3f-32a901ec6a07 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.338906] env[61986]: DEBUG nova.network.neutron [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Successfully updated port: b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.343943] env[61986]: DEBUG nova.scheduler.client.report [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.417921] env[61986]: DEBUG nova.network.neutron [-] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.545216] env[61986]: DEBUG nova.compute.manager [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Received event network-changed-b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1028.545216] env[61986]: DEBUG nova.compute.manager [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Refreshing instance network info cache due to event network-changed-b4336d9c-969c-4d2b-830f-c988c73c8410. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1028.545216] env[61986]: DEBUG oslo_concurrency.lockutils [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] Acquiring lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.545347] env[61986]: DEBUG oslo_concurrency.lockutils [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] Acquired lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.545607] env[61986]: DEBUG nova.network.neutron [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Refreshing network info cache for port b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1028.585637] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30bb28e7-2e98-4d02-b78d-ce3d2a4b038d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.596018] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d82d9d1-3aff-4869-85c7-1e54242c5938 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.620828] env[61986]: DEBUG nova.compute.manager [req-c9eaa1b6-d280-4214-829c-20e08a6346f1 req-4e6217ce-fb9e-42eb-aa18-623ec5c470f0 service nova] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Detach interface failed, port_id=74f613a6-57a3-4f00-b018-8c58f254200c, reason: Instance bda7750f-eec9-40d9-ace5-18d48234126e could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1028.715186] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551904} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.718348] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 645ecfee-3c0b-47c1-86c1-8ca21781438a/645ecfee-3c0b-47c1-86c1-8ca21781438a.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1028.718579] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.718836] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160307, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.719066] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2effae5f-30e0-42e0-b4d7-189234ebaab4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.725234] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1028.725234] env[61986]: value = "task-1160309" [ 1028.725234] env[61986]: _type = "Task" [ 1028.725234] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.732357] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160309, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.841859] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.848806] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.853121] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.917s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.853417] env[61986]: INFO nova.compute.claims [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.883172] env[61986]: INFO nova.scheduler.client.report [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Deleted allocations for instance aa12919e-3891-4da9-a280-4155137864e1 [ 1028.921213] env[61986]: INFO nova.compute.manager [-] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Took 2.05 seconds to deallocate network for instance. [ 1029.067648] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1029.103219] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.103464] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.103629] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.103819] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.103969] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.104137] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.104337] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.104500] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.104669] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.104832] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.105016] env[61986]: DEBUG nova.virt.hardware [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.105894] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315d74e2-b12a-45fa-8e1d-0dd6334fdb53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.109125] env[61986]: DEBUG nova.network.neutron [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1029.116280] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366a6d0d-0f52-4f4c-8004-b758731125f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.213952] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160307, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.235795] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160309, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066122} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.235795] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.235795] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8002367d-5de1-4b85-939e-9e361b474622 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.260518] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 645ecfee-3c0b-47c1-86c1-8ca21781438a/645ecfee-3c0b-47c1-86c1-8ca21781438a.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.261252] env[61986]: DEBUG nova.network.neutron [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.262803] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44f291c4-0cd8-4278-a645-f579874db799 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.285915] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1029.285915] env[61986]: value = "task-1160310" [ 1029.285915] env[61986]: _type = "Task" [ 1029.285915] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.294364] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160310, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.392819] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b89a8201-fea5-488d-ad5a-c1cffc8eb561 tempest-ServersAdminTestJSON-507632329 tempest-ServersAdminTestJSON-507632329-project-member] Lock "aa12919e-3891-4da9-a280-4155137864e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.667s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.429970] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.616157] env[61986]: DEBUG nova.network.neutron [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Successfully updated port: 85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.713740] env[61986]: DEBUG nova.network.neutron [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance_info_cache with network_info: [{"id": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "address": "fa:16:3e:c3:f2:46", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f71edfb-58", "ovs_interfaceid": "9f71edfb-58a5-43cb-9d35-62cda5bf87ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.721664] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160307, 'name': ReconfigVM_Task, 'duration_secs': 1.642715} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.725377] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 68626872-0875-4ff1-81ad-85bdff30ac94/68626872-0875-4ff1-81ad-85bdff30ac94.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.726715] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cfaa884f-031d-4f0a-8339-90019236b866 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.736749] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1029.736749] env[61986]: value = "task-1160311" [ 1029.736749] env[61986]: _type = "Task" [ 1029.736749] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.749745] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160311, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.781026] env[61986]: DEBUG oslo_concurrency.lockutils [req-7ce6829d-b968-43eb-ab9c-a7d8848d6c21 req-f42fda7d-c160-462d-8945-5d8b2acfc585 service nova] Releasing lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.781026] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.781265] env[61986]: DEBUG nova.network.neutron [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1029.799588] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160310, 'name': ReconfigVM_Task, 'duration_secs': 0.27286} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.799588] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 645ecfee-3c0b-47c1-86c1-8ca21781438a/645ecfee-3c0b-47c1-86c1-8ca21781438a.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.800088] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7cd2acf7-3e23-4b2f-b740-fd646e0f0eb0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.807034] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1029.807034] env[61986]: value = "task-1160312" [ 1029.807034] env[61986]: _type = "Task" [ 1029.807034] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.818089] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160312, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.058908] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b91d395-b56e-4a52-b606-184471911caa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.067330] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2f163f-a13d-43be-94b8-110585f24e64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.101134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a61e99-a174-4a77-be4c-4394e27b6e3f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.107993] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb1785d-867f-411d-9dab-e80ffb94f40d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.121900] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "refresh_cache-18893690-92d4-4391-a395-5cf2115397c3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.122066] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquired lock "refresh_cache-18893690-92d4-4391-a395-5cf2115397c3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.122216] env[61986]: DEBUG nova.network.neutron [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.123694] env[61986]: DEBUG nova.compute.provider_tree [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.143765] env[61986]: DEBUG nova.compute.manager [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Received event network-vif-plugged-85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.144077] env[61986]: DEBUG oslo_concurrency.lockutils [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] Acquiring lock "18893690-92d4-4391-a395-5cf2115397c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.144244] env[61986]: DEBUG oslo_concurrency.lockutils [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] Lock "18893690-92d4-4391-a395-5cf2115397c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.144379] env[61986]: DEBUG oslo_concurrency.lockutils [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] Lock "18893690-92d4-4391-a395-5cf2115397c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.144830] env[61986]: DEBUG nova.compute.manager [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] No waiting events found dispatching network-vif-plugged-85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.144830] env[61986]: WARNING nova.compute.manager [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Received unexpected event network-vif-plugged-85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 for instance with vm_state building and task_state spawning. [ 1030.144954] env[61986]: DEBUG nova.compute.manager [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Received event network-changed-85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.145188] env[61986]: DEBUG nova.compute.manager [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Refreshing instance network info cache due to event network-changed-85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1030.145367] env[61986]: DEBUG oslo_concurrency.lockutils [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] Acquiring lock "refresh_cache-18893690-92d4-4391-a395-5cf2115397c3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.217674] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-eb10ea87-8019-4f3b-af3f-32a901ec6a07" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.217674] env[61986]: DEBUG nova.objects.instance [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lazy-loading 'migration_context' on Instance uuid eb10ea87-8019-4f3b-af3f-32a901ec6a07 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.249774] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160311, 'name': Rename_Task, 'duration_secs': 0.130019} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.250040] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1030.250298] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa3d2d5a-a0b8-4fc1-ad20-695133cf1a44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.257160] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1030.257160] env[61986]: value = "task-1160313" [ 1030.257160] env[61986]: _type = "Task" [ 1030.257160] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.267041] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160313, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.316831] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160312, 'name': Rename_Task, 'duration_secs': 0.129544} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.317346] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1030.317819] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07fd9dc2-1402-46a6-97f5-9d5eef195bac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.324982] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1030.324982] env[61986]: value = "task-1160314" [ 1030.324982] env[61986]: _type = "Task" [ 1030.324982] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.333702] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160314, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.337691] env[61986]: DEBUG nova.network.neutron [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.540231] env[61986]: DEBUG nova.network.neutron [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updating instance_info_cache with network_info: [{"id": "b4336d9c-969c-4d2b-830f-c988c73c8410", "address": "fa:16:3e:a6:54:97", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4336d9c-96", "ovs_interfaceid": "b4336d9c-969c-4d2b-830f-c988c73c8410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.626809] env[61986]: DEBUG nova.scheduler.client.report [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.688509] env[61986]: DEBUG nova.network.neutron [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.719660] env[61986]: DEBUG nova.objects.base [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1030.723511] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1681436-3f81-4614-a701-95daa475a869 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.751321] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-633bf2f1-1334-45d1-ade3-275a5c9f749c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.758650] env[61986]: DEBUG oslo_vmware.api [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1030.758650] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d09b65-c2fd-0b7b-a10a-8ecda12e1e9d" [ 1030.758650] env[61986]: _type = "Task" [ 1030.758650] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.776226] env[61986]: DEBUG oslo_vmware.api [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160313, 'name': PowerOnVM_Task, 'duration_secs': 0.470912} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.776287] env[61986]: DEBUG oslo_vmware.api [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d09b65-c2fd-0b7b-a10a-8ecda12e1e9d, 'name': SearchDatastore_Task} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.776478] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.776676] env[61986]: INFO nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Took 6.38 seconds to spawn the instance on the hypervisor. [ 1030.776851] env[61986]: DEBUG nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.777186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.778117] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d745e3-7bc8-4eba-b9f8-6d32933685cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.837607] env[61986]: DEBUG oslo_vmware.api [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160314, 'name': PowerOnVM_Task, 'duration_secs': 0.439507} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.838335] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.838621] env[61986]: INFO nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Took 10.54 seconds to spawn the instance on the hypervisor. [ 1030.838863] env[61986]: DEBUG nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.840745] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deee1467-c3af-472b-adf4-bfc53e714fd0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.962406] env[61986]: DEBUG nova.network.neutron [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Updating instance_info_cache with network_info: [{"id": "85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54", "address": "fa:16:3e:c1:46:7c", "network": {"id": "2ff41b1a-bba0-4e4d-9e55-f475dfbbd3ad", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1735659882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86fcc2bbe80645b5a5cde8b7837b7575", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85b66ae5-e2", "ovs_interfaceid": "85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.045017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.045017] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance network_info: |[{"id": "b4336d9c-969c-4d2b-830f-c988c73c8410", "address": "fa:16:3e:a6:54:97", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4336d9c-96", "ovs_interfaceid": "b4336d9c-969c-4d2b-830f-c988c73c8410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.045017] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:54:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4336d9c-969c-4d2b-830f-c988c73c8410', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.051143] env[61986]: DEBUG oslo.service.loopingcall [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.051782] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1031.052138] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b342c12-19ad-406b-8a68-4b60eaa04162 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.074429] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.074429] env[61986]: value = "task-1160315" [ 1031.074429] env[61986]: _type = "Task" [ 1031.074429] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.084085] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160315, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.140145] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.140145] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.142667] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.770s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.142667] env[61986]: DEBUG nova.objects.instance [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'resources' on Instance uuid 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.302218] env[61986]: INFO nova.compute.manager [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Took 24.13 seconds to build instance. [ 1031.372846] env[61986]: INFO nova.compute.manager [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Took 25.57 seconds to build instance. [ 1031.465349] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Releasing lock "refresh_cache-18893690-92d4-4391-a395-5cf2115397c3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.468023] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Instance network_info: |[{"id": "85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54", "address": "fa:16:3e:c1:46:7c", "network": {"id": "2ff41b1a-bba0-4e4d-9e55-f475dfbbd3ad", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1735659882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86fcc2bbe80645b5a5cde8b7837b7575", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85b66ae5-e2", "ovs_interfaceid": "85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.468023] env[61986]: DEBUG oslo_concurrency.lockutils [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] Acquired lock "refresh_cache-18893690-92d4-4391-a395-5cf2115397c3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.468023] env[61986]: DEBUG nova.network.neutron [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Refreshing network info cache for port 85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.468023] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:46:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.480169] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Creating folder: Project (86fcc2bbe80645b5a5cde8b7837b7575). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1031.484346] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03704b48-fab0-4b96-b8f6-b7ce2419aeee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.497761] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Created folder: Project (86fcc2bbe80645b5a5cde8b7837b7575) in parent group-v252271. [ 1031.497992] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Creating folder: Instances. Parent ref: group-v252446. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1031.498266] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc4bfad7-8305-453d-bb04-f88702ff108b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.511416] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Created folder: Instances in parent group-v252446. [ 1031.511668] env[61986]: DEBUG oslo.service.loopingcall [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.511936] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1031.512096] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-705f2caa-7dbe-426b-82b0-df02bc672e98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.541022] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.541022] env[61986]: value = "task-1160318" [ 1031.541022] env[61986]: _type = "Task" [ 1031.541022] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.549992] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160318, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.585525] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160315, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.645568] env[61986]: DEBUG nova.compute.utils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.650778] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.650778] env[61986]: DEBUG nova.network.neutron [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1031.651178] env[61986]: DEBUG nova.objects.instance [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'numa_topology' on Instance uuid 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.709078] env[61986]: DEBUG nova.policy [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3481412ab54a0fb386a6d044a9ae7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c88c612fe00a4ad589db67f3016dace6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1031.809369] env[61986]: DEBUG oslo_concurrency.lockutils [None req-baeb0c94-bbf0-474b-9deb-c6537d9de16d tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "68626872-0875-4ff1-81ad-85bdff30ac94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.647s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.834060] env[61986]: DEBUG nova.network.neutron [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Updated VIF entry in instance network info cache for port 85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1031.834684] env[61986]: DEBUG nova.network.neutron [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Updating instance_info_cache with network_info: [{"id": "85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54", "address": "fa:16:3e:c1:46:7c", "network": {"id": "2ff41b1a-bba0-4e4d-9e55-f475dfbbd3ad", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1735659882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86fcc2bbe80645b5a5cde8b7837b7575", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85b66ae5-e2", "ovs_interfaceid": "85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.876035] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6bf7438-9cef-47ea-b88a-929a2b8c6442 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.086s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.053510] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160318, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.085711] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160315, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.155161] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.159924] env[61986]: DEBUG nova.objects.base [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Object Instance<221e10a3-da31-410c-80f8-4bcc2c515710> lazy-loaded attributes: resources,numa_topology {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1032.206091] env[61986]: DEBUG nova.network.neutron [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Successfully created port: 1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.340527] env[61986]: DEBUG oslo_concurrency.lockutils [req-647a8194-27c6-4f66-9e6f-ca55fb892893 req-903abac2-ed1b-4432-a459-0cffe6ca46ca service nova] Releasing lock "refresh_cache-18893690-92d4-4391-a395-5cf2115397c3" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.361919] env[61986]: DEBUG nova.compute.manager [None req-9d12a329-fb93-4962-bec3-b93f4b5f847e tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.362875] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e580e01-8f85-414f-9987-b17bac902ad0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.395250] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c298550f-5817-4c90-8018-df5ccc2f6766 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.403715] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2d0f71-a04c-4cc9-8969-14a5a20814f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.440441] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "645ecfee-3c0b-47c1-86c1-8ca21781438a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.440868] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.441196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.441196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.441368] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.443733] env[61986]: INFO nova.compute.manager [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Terminating instance [ 1032.445492] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0485e47-ad57-4402-8d90-eb178c795bd1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.448495] env[61986]: DEBUG nova.compute.manager [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.448621] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1032.449629] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6fff43-2740-4529-a2f8-b4be8fa1b567 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.459544] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052ca2b8-fec5-43b1-b729-2f3d82e0bfaf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.463219] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.463465] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-503a7a7e-06a0-4083-91d2-e0990d918876 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.476641] env[61986]: DEBUG nova.compute.provider_tree [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.480664] env[61986]: DEBUG oslo_vmware.api [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1032.480664] env[61986]: value = "task-1160319" [ 1032.480664] env[61986]: _type = "Task" [ 1032.480664] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.489589] env[61986]: DEBUG oslo_vmware.api [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.523144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "68626872-0875-4ff1-81ad-85bdff30ac94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.523144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "68626872-0875-4ff1-81ad-85bdff30ac94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.523144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "68626872-0875-4ff1-81ad-85bdff30ac94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.523144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "68626872-0875-4ff1-81ad-85bdff30ac94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.523144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "68626872-0875-4ff1-81ad-85bdff30ac94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.524400] env[61986]: INFO nova.compute.manager [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Terminating instance [ 1032.526307] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "refresh_cache-68626872-0875-4ff1-81ad-85bdff30ac94" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.526492] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquired lock "refresh_cache-68626872-0875-4ff1-81ad-85bdff30ac94" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.526681] env[61986]: DEBUG nova.network.neutron [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1032.553791] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160318, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.585793] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160315, 'name': CreateVM_Task, 'duration_secs': 1.503376} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.585979] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1032.586688] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.586871] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.587254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1032.587527] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20589210-7e0f-4f9c-85c3-057731d8b8f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.592051] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1032.592051] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52116d3a-fc7e-e091-240a-04bd09e574a5" [ 1032.592051] env[61986]: _type = "Task" [ 1032.592051] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.600444] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52116d3a-fc7e-e091-240a-04bd09e574a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.721879] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.721983] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.722200] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.722391] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.722570] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.726151] env[61986]: INFO nova.compute.manager [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Terminating instance [ 1032.728009] env[61986]: DEBUG nova.compute.manager [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.728219] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1032.729134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e142a10-7408-42e7-baa6-72cade0a3184 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.736671] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.736914] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4fb6baf-9a4b-4262-abc6-445abc1a23fc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.742713] env[61986]: DEBUG oslo_vmware.api [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1032.742713] env[61986]: value = "task-1160320" [ 1032.742713] env[61986]: _type = "Task" [ 1032.742713] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.752130] env[61986]: DEBUG oslo_vmware.api [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160320, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.876456] env[61986]: INFO nova.compute.manager [None req-9d12a329-fb93-4962-bec3-b93f4b5f847e tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] instance snapshotting [ 1032.877312] env[61986]: DEBUG nova.objects.instance [None req-9d12a329-fb93-4962-bec3-b93f4b5f847e tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lazy-loading 'flavor' on Instance uuid 68626872-0875-4ff1-81ad-85bdff30ac94 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.981614] env[61986]: DEBUG nova.scheduler.client.report [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.998238] env[61986]: DEBUG oslo_vmware.api [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160319, 'name': PowerOffVM_Task, 'duration_secs': 0.204868} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.998594] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.998827] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1032.999134] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94e3a668-18c6-41e2-bc40-e5a67e85761e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.052916] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160318, 'name': CreateVM_Task, 'duration_secs': 1.21533} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.053109] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1033.053757] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.054961] env[61986]: DEBUG nova.network.neutron [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.104335] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52116d3a-fc7e-e091-240a-04bd09e574a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011242} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.104335] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.104335] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.104614] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.104614] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.104703] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.104950] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.105283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.105511] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a45433f5-dfc5-4767-82ef-d41dc1bfa9e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.107528] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1db6781-5662-4f2f-9af3-97f13900d5a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.113236] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1033.113236] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52137c23-82e5-53b0-ce79-ad0da7ae3246" [ 1033.113236] env[61986]: _type = "Task" [ 1033.113236] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.117549] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.117761] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1033.119662] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29649570-b498-4de3-9318-d0b2caff576b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.121923] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1033.122132] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1033.122317] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Deleting the datastore file [datastore1] 645ecfee-3c0b-47c1-86c1-8ca21781438a {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.125293] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e437159-3e5a-4c03-acbc-7fc305a1cf62 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.126957] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52137c23-82e5-53b0-ce79-ad0da7ae3246, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.130327] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1033.130327] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525285a2-cfc5-3242-c232-f05b078e5265" [ 1033.130327] env[61986]: _type = "Task" [ 1033.130327] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.137606] env[61986]: DEBUG oslo_vmware.api [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for the task: (returnval){ [ 1033.137606] env[61986]: value = "task-1160322" [ 1033.137606] env[61986]: _type = "Task" [ 1033.137606] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.143874] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525285a2-cfc5-3242-c232-f05b078e5265, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.148737] env[61986]: DEBUG oslo_vmware.api [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.173264] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.180395] env[61986]: DEBUG nova.network.neutron [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.212682] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.212990] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.213325] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.213541] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.213691] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.213844] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.214077] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.214254] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.214429] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.214597] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.214774] env[61986]: DEBUG nova.virt.hardware [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.216782] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4a1948-734a-4053-ace6-b5e34f56e793 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.225612] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5b8b0d-55c8-462d-a751-32c3892b396e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.252664] env[61986]: DEBUG oslo_vmware.api [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160320, 'name': PowerOffVM_Task, 'duration_secs': 0.392795} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.252664] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1033.252664] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1033.253023] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e708e34-4388-4cdf-97a5-fddd4c6b8ffa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.310540] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1033.310817] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1033.310998] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleting the datastore file [datastore2] ee47cab9-5dfd-48ce-ba70-cb800d735b19 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.311302] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-206767b2-78a2-4aa6-bfee-33ace1bbf1c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.326822] env[61986]: DEBUG oslo_vmware.api [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for the task: (returnval){ [ 1033.326822] env[61986]: value = "task-1160324" [ 1033.326822] env[61986]: _type = "Task" [ 1033.326822] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.335309] env[61986]: DEBUG oslo_vmware.api [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.384777] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9191799-098b-4106-b09a-2fd388f2e453 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.403592] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7831f81-2ed8-4448-beb3-a13ec7bdf1f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.492184] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.351s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.494959] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.294s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.495289] env[61986]: DEBUG nova.objects.instance [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lazy-loading 'resources' on Instance uuid 45d80bab-0140-4473-9dc2-4a5481838aed {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.630927] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52137c23-82e5-53b0-ce79-ad0da7ae3246, 'name': SearchDatastore_Task, 'duration_secs': 0.023823} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.638715] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.639156] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.639524] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.648520] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525285a2-cfc5-3242-c232-f05b078e5265, 'name': SearchDatastore_Task, 'duration_secs': 0.011503} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.652349] env[61986]: DEBUG oslo_vmware.api [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Task: {'id': task-1160322, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259213} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.652702] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f1d424-d636-4664-b25d-66e09302a883 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.655055] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.658610] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1033.658610] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1033.658610] env[61986]: INFO nova.compute.manager [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1033.658610] env[61986]: DEBUG oslo.service.loopingcall [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.658610] env[61986]: DEBUG nova.compute.manager [-] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.658610] env[61986]: DEBUG nova.network.neutron [-] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1033.661631] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1033.661631] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5258ab6c-df4e-5207-f5ef-b1bd67023502" [ 1033.661631] env[61986]: _type = "Task" [ 1033.661631] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.669562] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5258ab6c-df4e-5207-f5ef-b1bd67023502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.685335] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Releasing lock "refresh_cache-68626872-0875-4ff1-81ad-85bdff30ac94" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.687769] env[61986]: DEBUG nova.compute.manager [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1033.687769] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1033.687769] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721c8886-0033-4f1e-a5a7-2cd0e4da968d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.695141] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1033.695141] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-acd7c702-9a77-400c-a412-ab006a33d705 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.701573] env[61986]: DEBUG oslo_vmware.api [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1033.701573] env[61986]: value = "task-1160325" [ 1033.701573] env[61986]: _type = "Task" [ 1033.701573] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.713942] env[61986]: DEBUG oslo_vmware.api [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.735527] env[61986]: DEBUG nova.compute.manager [req-888fdce6-4d72-4575-846f-b6d8b3cae7b7 req-fa87c6ca-4226-4040-8836-7e6bffb7f80e service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Received event network-vif-plugged-1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.735782] env[61986]: DEBUG oslo_concurrency.lockutils [req-888fdce6-4d72-4575-846f-b6d8b3cae7b7 req-fa87c6ca-4226-4040-8836-7e6bffb7f80e service nova] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.735968] env[61986]: DEBUG oslo_concurrency.lockutils [req-888fdce6-4d72-4575-846f-b6d8b3cae7b7 req-fa87c6ca-4226-4040-8836-7e6bffb7f80e service nova] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.736205] env[61986]: DEBUG oslo_concurrency.lockutils [req-888fdce6-4d72-4575-846f-b6d8b3cae7b7 req-fa87c6ca-4226-4040-8836-7e6bffb7f80e service nova] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.736325] env[61986]: DEBUG nova.compute.manager [req-888fdce6-4d72-4575-846f-b6d8b3cae7b7 req-fa87c6ca-4226-4040-8836-7e6bffb7f80e service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] No waiting events found dispatching network-vif-plugged-1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.736688] env[61986]: WARNING nova.compute.manager [req-888fdce6-4d72-4575-846f-b6d8b3cae7b7 req-fa87c6ca-4226-4040-8836-7e6bffb7f80e service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Received unexpected event network-vif-plugged-1a9aae37-5d10-441b-a606-1d2451e00510 for instance with vm_state building and task_state spawning. [ 1033.837679] env[61986]: DEBUG oslo_vmware.api [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Task: {'id': task-1160324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174208} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.837679] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.837859] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1033.838052] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1033.838270] env[61986]: INFO nova.compute.manager [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1033.838486] env[61986]: DEBUG oslo.service.loopingcall [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.838755] env[61986]: DEBUG nova.compute.manager [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.838850] env[61986]: DEBUG nova.network.neutron [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1033.865644] env[61986]: DEBUG nova.network.neutron [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Successfully updated port: 1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1033.917404] env[61986]: DEBUG nova.compute.manager [None req-9d12a329-fb93-4962-bec3-b93f4b5f847e tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Instance disappeared during snapshot {{(pid=61986) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 1033.972888] env[61986]: DEBUG nova.compute.manager [req-a6ad47a5-d9d8-4ec9-93fb-e4586b97a9cc req-a9a97182-748e-4e02-9888-8c3bd70e6d99 service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received event network-vif-deleted-25671e16-307f-48e8-b99f-891dea3c0a3a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.972888] env[61986]: INFO nova.compute.manager [req-a6ad47a5-d9d8-4ec9-93fb-e4586b97a9cc req-a9a97182-748e-4e02-9888-8c3bd70e6d99 service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Neutron deleted interface 25671e16-307f-48e8-b99f-891dea3c0a3a; detaching it from the instance and deleting it from the info cache [ 1033.973091] env[61986]: DEBUG nova.network.neutron [req-a6ad47a5-d9d8-4ec9-93fb-e4586b97a9cc req-a9a97182-748e-4e02-9888-8c3bd70e6d99 service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Updating instance_info_cache with network_info: [{"id": "c005e588-3e50-435a-849a-95a38ccac917", "address": "fa:16:3e:53:fa:ea", "network": {"id": "a2aa2876-c3fa-4902-ba57-50f984976a70", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1745244369", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c64e7f5b4b64d1988edf2a2fcbe32eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8459aaf-d6a8-46fb-ad14-464ac3104695", "external-id": "nsx-vlan-transportzone-46", "segmentation_id": 46, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc005e588-3e", "ovs_interfaceid": "c005e588-3e50-435a-849a-95a38ccac917", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.010627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f86bf5e-bd9c-4f78-a216-b0adeb0f5ed9 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 31.939s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.014411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 13.047s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.014411] env[61986]: INFO nova.compute.manager [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Unshelving [ 1034.065418] env[61986]: DEBUG nova.compute.manager [None req-9d12a329-fb93-4962-bec3-b93f4b5f847e tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Found 0 images (rotation: 2) {{(pid=61986) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1034.177834] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5258ab6c-df4e-5207-f5ef-b1bd67023502, 'name': SearchDatastore_Task, 'duration_secs': 0.017808} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.180509] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.180877] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] af31ba6f-0a3e-4a1c-af5e-617e8843839f/af31ba6f-0a3e-4a1c-af5e-617e8843839f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1034.181249] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.181492] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.181887] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-869320e5-157a-4708-950a-7d6c8beda919 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.184865] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-633bc685-f3d7-406e-91cf-5d0ca73b6abd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.194490] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1034.194490] env[61986]: value = "task-1160326" [ 1034.194490] env[61986]: _type = "Task" [ 1034.194490] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.197368] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.197368] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1034.202201] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ade44b21-6bd1-443c-bbb0-f17433c3cbf7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.213335] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160326, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.214741] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1034.214741] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5268ee17-3dc1-1182-85ad-0b4752dfde81" [ 1034.214741] env[61986]: _type = "Task" [ 1034.214741] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.223060] env[61986]: DEBUG oslo_vmware.api [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160325, 'name': PowerOffVM_Task, 'duration_secs': 0.118753} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.223888] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1034.224128] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1034.224415] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d11c7b7e-eb98-4733-b12a-d5a49905e4bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.229038] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5268ee17-3dc1-1182-85ad-0b4752dfde81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.234388] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca46a39-5b66-42b9-a919-433eb653b260 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.241772] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5934f63-f944-4814-b4fe-0078e4f7b30d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.250349] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1034.250579] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1034.250763] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Deleting the datastore file [datastore1] 68626872-0875-4ff1-81ad-85bdff30ac94 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.251474] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-283c098f-f65c-4a73-ba7e-f4790ba101f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.277726] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0080869f-3269-475f-9aeb-7d56f0d7247d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.281957] env[61986]: DEBUG oslo_vmware.api [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for the task: (returnval){ [ 1034.281957] env[61986]: value = "task-1160328" [ 1034.281957] env[61986]: _type = "Task" [ 1034.281957] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.288373] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6830c54-b852-48d7-baf1-769eeda49750 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.294913] env[61986]: DEBUG oslo_vmware.api [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160328, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.304241] env[61986]: DEBUG nova.compute.provider_tree [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.368823] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.369045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.369204] env[61986]: DEBUG nova.network.neutron [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.476259] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9515107f-776b-47e7-99b2-606562b73a60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.488154] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f20a95-e8e0-4ef2-ba83-e4022bc2962a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.519330] env[61986]: DEBUG nova.compute.manager [req-a6ad47a5-d9d8-4ec9-93fb-e4586b97a9cc req-a9a97182-748e-4e02-9888-8c3bd70e6d99 service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Detach interface failed, port_id=25671e16-307f-48e8-b99f-891dea3c0a3a, reason: Instance 645ecfee-3c0b-47c1-86c1-8ca21781438a could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1034.653460] env[61986]: DEBUG nova.network.neutron [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.687437] env[61986]: DEBUG nova.network.neutron [-] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.706504] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160326, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467791} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.706764] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] af31ba6f-0a3e-4a1c-af5e-617e8843839f/af31ba6f-0a3e-4a1c-af5e-617e8843839f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1034.706971] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.707229] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-384e082b-f5cf-4fd0-9195-5579276a0b41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.714872] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1034.714872] env[61986]: value = "task-1160329" [ 1034.714872] env[61986]: _type = "Task" [ 1034.714872] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.725844] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160329, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.729935] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5268ee17-3dc1-1182-85ad-0b4752dfde81, 'name': SearchDatastore_Task, 'duration_secs': 0.017199} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.730428] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02a0e6a5-380e-4088-b04b-3f50fdce92d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.735922] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1034.735922] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52abb13c-26b9-91e0-ec7a-a4da848f6a4b" [ 1034.735922] env[61986]: _type = "Task" [ 1034.735922] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.745447] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52abb13c-26b9-91e0-ec7a-a4da848f6a4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.792674] env[61986]: DEBUG oslo_vmware.api [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Task: {'id': task-1160328, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248795} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.792953] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.793153] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1034.793334] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1034.793514] env[61986]: INFO nova.compute.manager [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1034.793788] env[61986]: DEBUG oslo.service.loopingcall [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.794064] env[61986]: DEBUG nova.compute.manager [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1034.794175] env[61986]: DEBUG nova.network.neutron [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1034.807023] env[61986]: DEBUG nova.scheduler.client.report [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.811615] env[61986]: DEBUG nova.network.neutron [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1034.912437] env[61986]: DEBUG nova.network.neutron [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1035.025741] env[61986]: DEBUG nova.compute.utils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.124244] env[61986]: DEBUG nova.network.neutron [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.157586] env[61986]: INFO nova.compute.manager [-] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Took 1.32 seconds to deallocate network for instance. [ 1035.189997] env[61986]: INFO nova.compute.manager [-] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Took 1.53 seconds to deallocate network for instance. [ 1035.225989] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160329, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057883} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.225989] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.226421] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df696033-dc13-4346-998e-579d2923c2df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.250057] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] af31ba6f-0a3e-4a1c-af5e-617e8843839f/af31ba6f-0a3e-4a1c-af5e-617e8843839f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.253201] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f7fc5f9-1ad2-4bc6-b22b-33309ca4f2af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.273738] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52abb13c-26b9-91e0-ec7a-a4da848f6a4b, 'name': SearchDatastore_Task, 'duration_secs': 0.007513} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.276381] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.276381] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 18893690-92d4-4391-a395-5cf2115397c3/18893690-92d4-4391-a395-5cf2115397c3.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1035.276381] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1035.276381] env[61986]: value = "task-1160330" [ 1035.276381] env[61986]: _type = "Task" [ 1035.276381] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.276381] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-796060e0-020e-453d-8f92-e6fb4ee27e94 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.287386] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160330, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.288588] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1035.288588] env[61986]: value = "task-1160331" [ 1035.288588] env[61986]: _type = "Task" [ 1035.288588] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.295855] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.312223] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.314448] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.743s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.314687] env[61986]: DEBUG nova.objects.instance [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'resources' on Instance uuid db9050b8-897e-4d62-8dc2-3b334ab8afd4 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.315773] env[61986]: DEBUG nova.network.neutron [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.332210] env[61986]: INFO nova.scheduler.client.report [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Deleted allocations for instance 45d80bab-0140-4473-9dc2-4a5481838aed [ 1035.529095] env[61986]: INFO nova.virt.block_device [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Booting with volume 3f7d5322-8141-4aa7-813d-2ae4f033983e at /dev/sdb [ 1035.569346] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a739e651-978b-426a-b9d6-3ae1f023cb07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.578885] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f0e4c1-abc0-4066-9095-3fc36e2a5f88 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.605723] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eafa47ab-5bcd-49ce-a53d-2abe30815e13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.614315] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71eb6e7-b915-4a2c-8eb6-c1db3ac1c5c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.629049] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.629049] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Instance network_info: |[{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.630024] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:fb:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a9aae37-5d10-441b-a606-1d2451e00510', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.637758] env[61986]: DEBUG oslo.service.loopingcall [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.638421] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "fc3ae819-efa0-497e-8d2f-d17a53726056" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.638710] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "fc3ae819-efa0-497e-8d2f-d17a53726056" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.646090] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1035.646916] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2248b3a-7599-4893-b793-a88ce7a944f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.663045] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a509c9bc-8922-4396-9065-667466344e47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.666876] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.672521] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.672521] env[61986]: value = "task-1160332" [ 1035.672521] env[61986]: _type = "Task" [ 1035.672521] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.678892] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6ff443-db6d-45f6-8c59-c9c185b93bf0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.686981] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160332, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.693073] env[61986]: DEBUG nova.virt.block_device [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating existing volume attachment record: 136a5beb-c6fc-4d2e-bb53-de40e7b5f254 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1035.696581] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.767538] env[61986]: DEBUG nova.compute.manager [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Received event network-changed-1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.767776] env[61986]: DEBUG nova.compute.manager [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Refreshing instance network info cache due to event network-changed-1a9aae37-5d10-441b-a606-1d2451e00510. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1035.768011] env[61986]: DEBUG oslo_concurrency.lockutils [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.768166] env[61986]: DEBUG oslo_concurrency.lockutils [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.768337] env[61986]: DEBUG nova.network.neutron [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Refreshing network info cache for port 1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.789124] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160330, 'name': ReconfigVM_Task, 'duration_secs': 0.277264} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.789484] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Reconfigured VM instance instance-0000005d to attach disk [datastore1] af31ba6f-0a3e-4a1c-af5e-617e8843839f/af31ba6f-0a3e-4a1c-af5e-617e8843839f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.792813] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3044745-8101-4d63-a1b1-c75540ad5552 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.799637] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160331, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459996} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.800880] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 18893690-92d4-4391-a395-5cf2115397c3/18893690-92d4-4391-a395-5cf2115397c3.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1035.801111] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.801408] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1035.801408] env[61986]: value = "task-1160333" [ 1035.801408] env[61986]: _type = "Task" [ 1035.801408] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.801610] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a6190f9-af5e-4a38-b539-3b7d793b099c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.810809] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160333, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.812042] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1035.812042] env[61986]: value = "task-1160334" [ 1035.812042] env[61986]: _type = "Task" [ 1035.812042] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.823604] env[61986]: INFO nova.compute.manager [-] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Took 1.03 seconds to deallocate network for instance. [ 1035.823811] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160334, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.839735] env[61986]: DEBUG oslo_concurrency.lockutils [None req-10fccb11-2993-453d-b10e-289d7a647ac1 tempest-ServerShowV247Test-266788089 tempest-ServerShowV247Test-266788089-project-member] Lock "45d80bab-0140-4473-9dc2-4a5481838aed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.375s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.023706] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81683efc-b34d-4122-9c8c-545849440464 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.032185] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f33ffb-a7bf-4dfc-b782-7be2a8267034 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.065014] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87825927-b6b5-4bf6-82f2-cff54bf0b58f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.108055] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2f8ab5-b298-428d-9072-079d5a5d365e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.108055] env[61986]: DEBUG nova.compute.provider_tree [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.149329] env[61986]: DEBUG nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.193020] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160332, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.315858] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160333, 'name': Rename_Task, 'duration_secs': 0.132496} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.321376] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1036.321867] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3935971-6380-425d-b6b3-085b3ed0ecb5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.328766] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160334, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06488} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.331212] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.334055] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.334438] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1036.334438] env[61986]: value = "task-1160338" [ 1036.334438] env[61986]: _type = "Task" [ 1036.334438] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.335295] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36100690-2885-4ab0-9003-a4926a4f72eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.369396] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 18893690-92d4-4391-a395-5cf2115397c3/18893690-92d4-4391-a395-5cf2115397c3.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.374343] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cf9bf39-3d5e-4fd2-8d2f-ae282c81702a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.390487] env[61986]: DEBUG nova.compute.manager [req-7739b1d7-6e6c-48f6-9aca-8870ff56f41f req-b5a34319-b414-4076-a77e-245cbf5f4159 service nova] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Received event network-vif-deleted-c005e588-3e50-435a-849a-95a38ccac917 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.391334] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160338, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.397026] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1036.397026] env[61986]: value = "task-1160339" [ 1036.397026] env[61986]: _type = "Task" [ 1036.397026] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.406959] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160339, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.590379] env[61986]: DEBUG nova.scheduler.client.report [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.648718] env[61986]: DEBUG nova.network.neutron [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updated VIF entry in instance network info cache for port 1a9aae37-5d10-441b-a606-1d2451e00510. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.649155] env[61986]: DEBUG nova.network.neutron [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.674676] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.687554] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160332, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.848930] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160338, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.906905] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160339, 'name': ReconfigVM_Task, 'duration_secs': 0.444621} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.907222] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 18893690-92d4-4391-a395-5cf2115397c3/18893690-92d4-4391-a395-5cf2115397c3.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.907861] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6c2bf15-8ebf-45c9-9426-7f9776807d39 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.914211] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1036.914211] env[61986]: value = "task-1160340" [ 1036.914211] env[61986]: _type = "Task" [ 1036.914211] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.923742] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160340, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.096031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.098313] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.381s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.098502] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.098707] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1037.099052] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.669s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.099265] env[61986]: DEBUG nova.objects.instance [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lazy-loading 'resources' on Instance uuid bda7750f-eec9-40d9-ace5-18d48234126e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.100968] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9138d4ab-e2fc-4495-adc5-7248ce04c58b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.109489] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b265dfe0-0dc5-48e2-9488-ebc6828e0dd4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.124487] env[61986]: INFO nova.scheduler.client.report [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocations for instance db9050b8-897e-4d62-8dc2-3b334ab8afd4 [ 1037.125866] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d158825a-bf8c-43e4-814f-f52bca3b6381 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.135208] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dc30f6-46f8-4851-bf8c-5c8dfc1b2440 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.164070] env[61986]: DEBUG oslo_concurrency.lockutils [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.164336] env[61986]: DEBUG nova.compute.manager [req-2fac86a3-fb62-4aee-b8a9-d9a07db2d765 req-286a7898-241c-4ea7-a05b-74d0354b5f33 service nova] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Received event network-vif-deleted-78729401-474d-4e4f-8159-aff2e908be0b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1037.165100] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179099MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1037.165246] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.188449] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160332, 'name': CreateVM_Task, 'duration_secs': 1.365943} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.188615] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1037.189287] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.189456] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.189769] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.190030] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f518474-8b88-4634-9c37-06b529d4bb60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.194649] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1037.194649] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52583fd7-9e20-c02d-4525-75b443e56191" [ 1037.194649] env[61986]: _type = "Task" [ 1037.194649] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.201865] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52583fd7-9e20-c02d-4525-75b443e56191, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.348496] env[61986]: DEBUG oslo_vmware.api [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160338, 'name': PowerOnVM_Task, 'duration_secs': 0.583395} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.348830] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1037.349090] env[61986]: INFO nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Took 10.64 seconds to spawn the instance on the hypervisor. [ 1037.349254] env[61986]: DEBUG nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1037.350298] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e8d1dd-5814-4f80-bbf2-3d3929693602 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.426641] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160340, 'name': Rename_Task, 'duration_secs': 0.463648} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.426898] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1037.427183] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cca1183-0db9-4e4f-a482-b2770f868527 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.433676] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1037.433676] env[61986]: value = "task-1160341" [ 1037.433676] env[61986]: _type = "Task" [ 1037.433676] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.441415] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160341, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.639587] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d7e555be-58b3-4a75-9c5f-e2168ca2ef5a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "db9050b8-897e-4d62-8dc2-3b334ab8afd4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.021s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.712807] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52583fd7-9e20-c02d-4525-75b443e56191, 'name': SearchDatastore_Task, 'duration_secs': 0.009047} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.713325] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.713720] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.714137] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.714411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.714727] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.719817] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bd55c6a-a794-4c77-9c26-3917b8f808ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.732060] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.732503] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1037.737228] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cfe81df-f114-4569-83c5-5e5314515dc4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.745080] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1037.745080] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52488077-8a0f-a843-a5fe-c252cbab6551" [ 1037.745080] env[61986]: _type = "Task" [ 1037.745080] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.756242] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52488077-8a0f-a843-a5fe-c252cbab6551, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.800807] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf92cea-0a43-4fb8-aca8-d946f6db4f9c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.811796] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319f6f78-3020-4684-b6a4-591be737e53a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.845298] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c460e1-5d20-4297-b7c7-bf46b9c78e31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.853901] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9fc303-9daa-4582-b0ab-ac2c975b1fcc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.874206] env[61986]: DEBUG nova.compute.provider_tree [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.877352] env[61986]: INFO nova.compute.manager [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Took 29.29 seconds to build instance. [ 1037.946230] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.257234] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52488077-8a0f-a843-a5fe-c252cbab6551, 'name': SearchDatastore_Task, 'duration_secs': 0.009019} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.258126] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10f47c29-4d46-4abf-b645-293aea2fffee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.264525] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1038.264525] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52839a9f-05b4-272d-a9ab-885d3a73bf1c" [ 1038.264525] env[61986]: _type = "Task" [ 1038.264525] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.273956] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52839a9f-05b4-272d-a9ab-885d3a73bf1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.378684] env[61986]: DEBUG nova.scheduler.client.report [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.384665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2bf9b548-da35-4a4c-bf01-7f32a6742707 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.813s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.445524] env[61986]: DEBUG oslo_vmware.api [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160341, 'name': PowerOnVM_Task, 'duration_secs': 0.868768} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.445924] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.446161] env[61986]: INFO nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Took 9.38 seconds to spawn the instance on the hypervisor. [ 1038.446348] env[61986]: DEBUG nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.447202] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967c47ff-a240-440b-8766-e7ace8603cae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.778077] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52839a9f-05b4-272d-a9ab-885d3a73bf1c, 'name': SearchDatastore_Task, 'duration_secs': 0.012267} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.778381] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.778652] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1038.778964] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3b709a9-5362-4017-a657-b1f380e74cdc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.793200] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1038.793200] env[61986]: value = "task-1160343" [ 1038.793200] env[61986]: _type = "Task" [ 1038.793200] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.802421] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160343, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.824554] env[61986]: DEBUG nova.compute.manager [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Received event network-changed-b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.824749] env[61986]: DEBUG nova.compute.manager [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Refreshing instance network info cache due to event network-changed-b4336d9c-969c-4d2b-830f-c988c73c8410. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1038.824976] env[61986]: DEBUG oslo_concurrency.lockutils [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] Acquiring lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.826399] env[61986]: DEBUG oslo_concurrency.lockutils [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] Acquired lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.826606] env[61986]: DEBUG nova.network.neutron [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Refreshing network info cache for port b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1038.888244] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.789s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.892343] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.113s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.925487] env[61986]: INFO nova.scheduler.client.report [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted allocations for instance bda7750f-eec9-40d9-ace5-18d48234126e [ 1038.968393] env[61986]: INFO nova.compute.manager [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Took 28.09 seconds to build instance. [ 1039.304510] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160343, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.426911] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.427284] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.437124] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a089395f-f7dc-4a82-b609-353cdd269b6b tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bda7750f-eec9-40d9-ace5-18d48234126e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.704s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.472268] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1912502b-1657-4e1a-bc9e-009822c0d217 tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "18893690-92d4-4391-a395-5cf2115397c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.604s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.585124] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.585419] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.585673] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.585889] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.586132] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.588556] env[61986]: INFO nova.compute.manager [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Terminating instance [ 1039.590758] env[61986]: DEBUG nova.compute.manager [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1039.590967] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1039.592210] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c78e5be-5031-4ae1-9751-ac1092984b89 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.605114] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1039.605378] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2252e49-340f-43d9-9e58-c8e568fa6668 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.612426] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7fcfbf-af1e-42f6-ab07-eb398706e006 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.616221] env[61986]: DEBUG oslo_vmware.api [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 1039.616221] env[61986]: value = "task-1160344" [ 1039.616221] env[61986]: _type = "Task" [ 1039.616221] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.622969] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2a1167-cb8d-43d0-9e86-e0ee08e604e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.629332] env[61986]: DEBUG oslo_vmware.api [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160344, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.660976] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c418f201-8aa8-4fa6-b746-8f8ad4b00326 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.675438] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b927fb8-501b-48db-811b-add954045b70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.692361] env[61986]: DEBUG nova.compute.provider_tree [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.753592] env[61986]: DEBUG nova.network.neutron [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updated VIF entry in instance network info cache for port b4336d9c-969c-4d2b-830f-c988c73c8410. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1039.754236] env[61986]: DEBUG nova.network.neutron [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updating instance_info_cache with network_info: [{"id": "b4336d9c-969c-4d2b-830f-c988c73c8410", "address": "fa:16:3e:a6:54:97", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4336d9c-96", "ovs_interfaceid": "b4336d9c-969c-4d2b-830f-c988c73c8410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.804984] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160343, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.930560] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1040.127084] env[61986]: DEBUG oslo_vmware.api [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160344, 'name': PowerOffVM_Task, 'duration_secs': 0.273055} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.127361] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.127533] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.127794] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0cdeaa7-42a7-434a-9969-29b0aeed5df5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.170201] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "18893690-92d4-4391-a395-5cf2115397c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.170561] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "18893690-92d4-4391-a395-5cf2115397c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.170878] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "18893690-92d4-4391-a395-5cf2115397c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.171107] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "18893690-92d4-4391-a395-5cf2115397c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.171296] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "18893690-92d4-4391-a395-5cf2115397c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.173513] env[61986]: INFO nova.compute.manager [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Terminating instance [ 1040.175369] env[61986]: DEBUG nova.compute.manager [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.175584] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.176430] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc4f64a-b673-46ef-8cf7-2602db2bf53e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.184082] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.184432] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-655b6fbc-4057-4816-8358-44f4c46d4491 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.193296] env[61986]: DEBUG oslo_vmware.api [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1040.193296] env[61986]: value = "task-1160346" [ 1040.193296] env[61986]: _type = "Task" [ 1040.193296] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.197483] env[61986]: DEBUG nova.scheduler.client.report [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.203210] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.203473] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.203610] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleting the datastore file [datastore2] bc93bdd5-256e-4900-a9ae-609b0a7d402d {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.203903] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-802632f7-ffa1-4b48-a79f-a1ac04bdfe32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.209126] env[61986]: DEBUG oslo_vmware.api [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160346, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.213685] env[61986]: DEBUG oslo_vmware.api [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for the task: (returnval){ [ 1040.213685] env[61986]: value = "task-1160347" [ 1040.213685] env[61986]: _type = "Task" [ 1040.213685] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.223147] env[61986]: DEBUG oslo_vmware.api [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160347, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.257062] env[61986]: DEBUG oslo_concurrency.lockutils [req-4f1e9f89-aa18-4086-afcb-5944a30d2afa req-a9df65b3-5180-416c-9ddb-5f3266acb0b9 service nova] Releasing lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.305639] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160343, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.452855] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.707243] env[61986]: DEBUG oslo_vmware.api [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160346, 'name': PowerOffVM_Task, 'duration_secs': 0.179821} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.707895] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.708104] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.708367] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-beae5b46-8b06-460b-a51b-cc09f85d9c6c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.723201] env[61986]: DEBUG oslo_vmware.api [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Task: {'id': task-1160347, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169733} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.724043] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1040.724232] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1040.724414] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1040.724591] env[61986]: INFO nova.compute.manager [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1040.724826] env[61986]: DEBUG oslo.service.loopingcall [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.725025] env[61986]: DEBUG nova.compute.manager [-] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1040.725123] env[61986]: DEBUG nova.network.neutron [-] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1040.791122] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.791357] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.792334] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Deleting the datastore file [datastore1] 18893690-92d4-4391-a395-5cf2115397c3 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.792414] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea40cad8-0953-436f-be33-9602c6abdfda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.799768] env[61986]: DEBUG oslo_vmware.api [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for the task: (returnval){ [ 1040.799768] env[61986]: value = "task-1160349" [ 1040.799768] env[61986]: _type = "Task" [ 1040.799768] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.806926] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160343, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.806548} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.807482] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1040.807701] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1040.807970] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b77de909-265d-4643-8336-ccb8ca34b6b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.812581] env[61986]: DEBUG oslo_vmware.api [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.816874] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1040.816874] env[61986]: value = "task-1160350" [ 1040.816874] env[61986]: _type = "Task" [ 1040.816874] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.824637] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160350, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.186336] env[61986]: DEBUG nova.compute.manager [req-aaac8a95-938f-443f-9010-78cd61aa9c3a req-b7b34430-1cf8-425b-85a6-37da8abc94a7 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Received event network-vif-deleted-51b89bbc-3efd-4e51-b21b-bfce45546eae {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1041.186549] env[61986]: INFO nova.compute.manager [req-aaac8a95-938f-443f-9010-78cd61aa9c3a req-b7b34430-1cf8-425b-85a6-37da8abc94a7 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Neutron deleted interface 51b89bbc-3efd-4e51-b21b-bfce45546eae; detaching it from the instance and deleting it from the info cache [ 1041.186727] env[61986]: DEBUG nova.network.neutron [req-aaac8a95-938f-443f-9010-78cd61aa9c3a req-b7b34430-1cf8-425b-85a6-37da8abc94a7 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.209444] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.319s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.214067] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.547s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.214445] env[61986]: DEBUG nova.objects.instance [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lazy-loading 'resources' on Instance uuid ee47cab9-5dfd-48ce-ba70-cb800d735b19 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.308847] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.318399] env[61986]: DEBUG oslo_vmware.api [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Task: {'id': task-1160349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129808} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.323502] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.323855] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.324172] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.324474] env[61986]: INFO nova.compute.manager [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1041.324854] env[61986]: DEBUG oslo.service.loopingcall [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.325192] env[61986]: DEBUG nova.compute.manager [-] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.325364] env[61986]: DEBUG nova.network.neutron [-] [instance: 18893690-92d4-4391-a395-5cf2115397c3] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1041.334846] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160350, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158773} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.335260] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1041.336517] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1742f2b-01bf-415a-9472-5307d57b9b65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.364144] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.366506] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ccce3e2-48f3-4f75-87cd-419bcef51f16 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.385998] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1041.385998] env[61986]: value = "task-1160351" [ 1041.385998] env[61986]: _type = "Task" [ 1041.385998] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.394025] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.581072] env[61986]: DEBUG nova.network.neutron [-] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.688941] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9ed3d7f-523e-4823-9b3e-5a0d4c1772f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.698351] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83a3f70-cf75-487f-9965-c8fccb3d3568 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.728577] env[61986]: DEBUG nova.compute.manager [req-aaac8a95-938f-443f-9010-78cd61aa9c3a req-b7b34430-1cf8-425b-85a6-37da8abc94a7 service nova] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Detach interface failed, port_id=51b89bbc-3efd-4e51-b21b-bfce45546eae, reason: Instance bc93bdd5-256e-4900-a9ae-609b0a7d402d could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1041.775649] env[61986]: INFO nova.scheduler.client.report [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocation for migration a54e6fce-6853-4631-9c9f-715c61d45e67 [ 1041.855754] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d2fbfa-299b-4a34-b371-8fd1cef4bfe5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.863153] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654319dc-5f40-4d11-9d5e-d9de045c45b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.894487] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c368a8cd-e8c1-4e8e-95fc-2d8b0682c009 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.904128] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b6c851-e97d-4450-bd54-a5f0e16b8d22 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.907980] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160351, 'name': ReconfigVM_Task, 'duration_secs': 0.273937} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.908293] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to attach disk [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.909167] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87fb812d-2b57-4389-b250-106e7081532e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.918529] env[61986]: DEBUG nova.compute.provider_tree [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.924238] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1041.924238] env[61986]: value = "task-1160352" [ 1041.924238] env[61986]: _type = "Task" [ 1041.924238] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.934824] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160352, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.084162] env[61986]: INFO nova.compute.manager [-] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Took 1.36 seconds to deallocate network for instance. [ 1042.103917] env[61986]: DEBUG nova.network.neutron [-] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.281790] env[61986]: DEBUG oslo_concurrency.lockutils [None req-751f29fc-287e-4bc8-ac1c-66972ab6f214 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.599s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.422983] env[61986]: DEBUG nova.scheduler.client.report [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.436070] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160352, 'name': Rename_Task, 'duration_secs': 0.133629} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.436378] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1042.436645] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3cf7227-0f2a-405a-836b-ffbf66a5f0ca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.443538] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1042.443538] env[61986]: value = "task-1160353" [ 1042.443538] env[61986]: _type = "Task" [ 1042.443538] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.452772] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160353, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.592208] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.606810] env[61986]: INFO nova.compute.manager [-] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Took 1.28 seconds to deallocate network for instance. [ 1042.930993] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.933479] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.237s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.933828] env[61986]: DEBUG nova.objects.instance [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lazy-loading 'resources' on Instance uuid 645ecfee-3c0b-47c1-86c1-8ca21781438a {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.952852] env[61986]: DEBUG oslo_vmware.api [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160353, 'name': PowerOnVM_Task, 'duration_secs': 0.456244} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.953774] env[61986]: INFO nova.scheduler.client.report [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Deleted allocations for instance ee47cab9-5dfd-48ce-ba70-cb800d735b19 [ 1042.954753] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.954963] env[61986]: INFO nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Took 9.78 seconds to spawn the instance on the hypervisor. [ 1042.955166] env[61986]: DEBUG nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.957969] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af55de7b-6b1d-45e2-a31d-b4eaea4a43a7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.016922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.017302] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.017535] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.017726] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.017903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.019961] env[61986]: INFO nova.compute.manager [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Terminating instance [ 1043.021772] env[61986]: DEBUG nova.compute.manager [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1043.021969] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1043.022809] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42f579a-c7ca-4a58-ae4a-85c3ba2ccb60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.030450] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1043.030718] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e22af146-0313-4296-8200-42373f66d224 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.037483] env[61986]: DEBUG oslo_vmware.api [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1043.037483] env[61986]: value = "task-1160354" [ 1043.037483] env[61986]: _type = "Task" [ 1043.037483] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.054251] env[61986]: DEBUG oslo_vmware.api [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.112903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.214089] env[61986]: DEBUG nova.compute.manager [req-891079dd-32d2-44ae-b0fb-c481bc8bfc28 req-9ea5054b-a130-405e-a2a7-8e15f01ad2ca service nova] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Received event network-vif-deleted-85b66ae5-e22f-4be7-a3b8-a84cd2f8bc54 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1043.467477] env[61986]: DEBUG oslo_concurrency.lockutils [None req-edc34e0c-75ab-462d-95d4-ed358b6e7f71 tempest-ServersNegativeTestJSON-2114299282 tempest-ServersNegativeTestJSON-2114299282-project-member] Lock "ee47cab9-5dfd-48ce-ba70-cb800d735b19" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.745s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.477849] env[61986]: INFO nova.compute.manager [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Took 24.56 seconds to build instance. [ 1043.518158] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "c3bcd1ba-a3df-4b81-9743-431235536642" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.518401] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "c3bcd1ba-a3df-4b81-9743-431235536642" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.547078] env[61986]: DEBUG oslo_vmware.api [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160354, 'name': PowerOffVM_Task, 'duration_secs': 0.180802} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.550103] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1043.550103] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1043.550246] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f428d13-c5d6-42f3-a85b-a146c4f8a0f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.599370] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41426e06-1b84-4189-84dd-5ccd96caefa1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.607166] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27df50f7-42f8-48ef-b594-8334b9eb0da0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.611799] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1043.612036] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1043.612229] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleting the datastore file [datastore2] eb10ea87-8019-4f3b-af3f-32a901ec6a07 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.612861] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5b71d3a-a3c6-4285-a861-32d1008e5ae3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.641308] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250127ed-773c-489d-a03c-056a8bf9be21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.644209] env[61986]: DEBUG oslo_vmware.api [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1043.644209] env[61986]: value = "task-1160356" [ 1043.644209] env[61986]: _type = "Task" [ 1043.644209] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.651352] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f4a1dc-8d0b-4f70-8082-9f84c1c1c338 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.658705] env[61986]: DEBUG oslo_vmware.api [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160356, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.668427] env[61986]: DEBUG nova.compute.provider_tree [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.980285] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a435c77f-91bc-48b7-990e-949a49f42e19 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.067s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.022989] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1044.154073] env[61986]: DEBUG oslo_vmware.api [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160356, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170182} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.154441] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.154731] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1044.154809] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1044.154987] env[61986]: INFO nova.compute.manager [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1044.155253] env[61986]: DEBUG oslo.service.loopingcall [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.155499] env[61986]: DEBUG nova.compute.manager [-] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.155574] env[61986]: DEBUG nova.network.neutron [-] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1044.172639] env[61986]: DEBUG nova.scheduler.client.report [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.546559] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.676923] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.679547] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.346s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.679774] env[61986]: DEBUG nova.objects.instance [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lazy-loading 'resources' on Instance uuid 68626872-0875-4ff1-81ad-85bdff30ac94 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.708975] env[61986]: INFO nova.scheduler.client.report [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Deleted allocations for instance 645ecfee-3c0b-47c1-86c1-8ca21781438a [ 1044.946152] env[61986]: DEBUG nova.network.neutron [-] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.218707] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924a8d76-a604-4b28-9fd9-732ad463b666 tempest-ServersTestMultiNic-83405650 tempest-ServersTestMultiNic-83405650-project-member] Lock "645ecfee-3c0b-47c1-86c1-8ca21781438a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.778s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.242460] env[61986]: DEBUG nova.compute.manager [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Received event network-changed-1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1045.242581] env[61986]: DEBUG nova.compute.manager [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Refreshing instance network info cache due to event network-changed-1a9aae37-5d10-441b-a606-1d2451e00510. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1045.243064] env[61986]: DEBUG oslo_concurrency.lockutils [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.243064] env[61986]: DEBUG oslo_concurrency.lockutils [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.243302] env[61986]: DEBUG nova.network.neutron [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Refreshing network info cache for port 1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.323263] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8c490f-8055-4f23-9e30-51a3805ea1b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.331346] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5ccc43-6755-4412-a2f2-bbee2a8f08d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.363808] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdae8a5-078a-4c4d-9a30-7d03b6e76b20 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.372129] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bc1697-5b92-4e28-9d82-906cfeb62cc9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.387480] env[61986]: DEBUG nova.compute.provider_tree [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.449133] env[61986]: INFO nova.compute.manager [-] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Took 1.29 seconds to deallocate network for instance. [ 1045.890333] env[61986]: DEBUG nova.scheduler.client.report [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.957380] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.049188] env[61986]: DEBUG nova.network.neutron [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updated VIF entry in instance network info cache for port 1a9aae37-5d10-441b-a606-1d2451e00510. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.049845] env[61986]: DEBUG nova.network.neutron [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.395891] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.398945] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.724s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.399722] env[61986]: INFO nova.compute.claims [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.431459] env[61986]: INFO nova.scheduler.client.report [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Deleted allocations for instance 68626872-0875-4ff1-81ad-85bdff30ac94 [ 1046.553296] env[61986]: DEBUG oslo_concurrency.lockutils [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.553618] env[61986]: DEBUG nova.compute.manager [req-595acc8d-fd41-4b83-935e-d3592ff232e7 req-09f3ecc7-809f-4c43-b105-0dda17b351d6 service nova] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Received event network-vif-deleted-9f71edfb-58a5-43cb-9d35-62cda5bf87ad {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.938793] env[61986]: DEBUG oslo_concurrency.lockutils [None req-41b0e3ca-5c8d-44ec-9638-d0fcab36a209 tempest-ServersAaction247Test-800167762 tempest-ServersAaction247Test-800167762-project-member] Lock "68626872-0875-4ff1-81ad-85bdff30ac94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.417s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.550015] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0b17ec-fb75-4ef8-9bd5-cc7c4c570031 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.558753] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5cabab-74ce-4dd8-b944-a82a1b59aff6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.592099] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805187ae-b348-4f4a-baee-d81df5a2d76a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.599542] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aaedb71-591f-45ff-bced-c09c7566a14a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.612954] env[61986]: DEBUG nova.compute.provider_tree [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.116562] env[61986]: DEBUG nova.scheduler.client.report [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.621880] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.622648] env[61986]: DEBUG nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1048.624951] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.460s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.130846] env[61986]: DEBUG nova.compute.utils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1049.139098] env[61986]: DEBUG nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Not allocating networking since 'none' was specified. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1049.640229] env[61986]: DEBUG nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1049.677669] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance bc93bdd5-256e-4900-a9ae-609b0a7d402d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1049.677669] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance eb10ea87-8019-4f3b-af3f-32a901ec6a07 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1049.677669] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance af31ba6f-0a3e-4a1c-af5e-617e8843839f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1049.677669] env[61986]: WARNING nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 18893690-92d4-4391-a395-5cf2115397c3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1049.677669] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance e825bf09-7f6c-45db-b6cf-9f2dabb98677 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.180438] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 221e10a3-da31-410c-80f8-4bcc2c515710 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1050.180737] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance fc3ae819-efa0-497e-8d2f-d17a53726056 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.656198] env[61986]: DEBUG nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1050.681917] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.682231] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.682392] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.682578] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.682727] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.682882] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.683120] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.683289] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.683470] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.683639] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.683816] env[61986]: DEBUG nova.virt.hardware [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.684533] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9efa4214-3954-49ef-bc7b-6daba8ba07f2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1050.686297] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5e1304-53c3-43fb-b2bb-9e4da6580ed9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.694709] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86be34c-479a-4d95-94b6-e73cbedf911a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.710580] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.716721] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Creating folder: Project (0ad474b464a149a8a752e0593b306a68). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1050.717265] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3470c392-575e-431c-9c2f-19268919d756 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.727882] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Created folder: Project (0ad474b464a149a8a752e0593b306a68) in parent group-v252271. [ 1050.728122] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Creating folder: Instances. Parent ref: group-v252452. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1050.728380] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-174843d4-6513-409d-931b-0b300c33bd09 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.737271] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Created folder: Instances in parent group-v252452. [ 1050.737525] env[61986]: DEBUG oslo.service.loopingcall [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.737728] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1050.737940] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd9655de-5c4d-4122-bb20-bebeee197205 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.754513] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.754513] env[61986]: value = "task-1160359" [ 1050.754513] env[61986]: _type = "Task" [ 1050.754513] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.762095] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160359, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.132044] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.132297] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.190259] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance c3bcd1ba-a3df-4b81-9743-431235536642 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1051.190651] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1051.190732] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1051.267121] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160359, 'name': CreateVM_Task, 'duration_secs': 0.253013} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.267121] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1051.267568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.268336] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.268336] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1051.268336] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24a955fe-c798-4728-8ff2-37d1bb6695f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.276279] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1051.276279] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5298d908-c4c3-f613-7450-fa3dce8cd933" [ 1051.276279] env[61986]: _type = "Task" [ 1051.276279] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.284936] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5298d908-c4c3-f613-7450-fa3dce8cd933, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.329487] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82bf157-3169-427b-8261-7a1e6f39afb1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.338195] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e23d56d-6878-49d1-873a-6564258b7f9d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.368586] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c82438a-dbfe-4f69-a621-29992ba7ad85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.375967] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af78fe1-8c0b-4c59-9830-d8d8a3e60874 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.394687] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.637924] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.792250] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5298d908-c4c3-f613-7450-fa3dce8cd933, 'name': SearchDatastore_Task, 'duration_secs': 0.010158} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.792618] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.792873] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.793131] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.793372] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.793687] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.794034] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1565b7af-898a-4b20-9d09-bccbcea89763 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.803219] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.803349] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1051.805328] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9871b5b3-3138-4fe2-a3db-6117fdebcb48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.811058] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1051.811058] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c1b0d0-9205-26c8-7e7d-0b210cc61920" [ 1051.811058] env[61986]: _type = "Task" [ 1051.811058] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.818976] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c1b0d0-9205-26c8-7e7d-0b210cc61920, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.897372] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.163186] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.321997] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c1b0d0-9205-26c8-7e7d-0b210cc61920, 'name': SearchDatastore_Task, 'duration_secs': 0.040286} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.322765] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ae304a-aea9-4ff8-bba2-bcadd8bee085 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.327628] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1052.327628] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d86a55-af36-61ee-7883-378297fbcf7c" [ 1052.327628] env[61986]: _type = "Task" [ 1052.327628] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.334816] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d86a55-af36-61ee-7883-378297fbcf7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.401993] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1052.402244] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.777s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.402527] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.950s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.404117] env[61986]: INFO nova.compute.claims [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.838098] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d86a55-af36-61ee-7883-378297fbcf7c, 'name': SearchDatastore_Task, 'duration_secs': 0.031344} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.838098] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.838318] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1052.838575] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80806b49-e79b-4ee7-8939-29761ef7bd0e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.844319] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1052.844319] env[61986]: value = "task-1160360" [ 1052.844319] env[61986]: _type = "Task" [ 1052.844319] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.851445] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160360, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.354202] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160360, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439181} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.354519] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1053.354645] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.354901] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6bcd40fc-a0f0-4b21-95e9-aed4838d50f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.362334] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1053.362334] env[61986]: value = "task-1160361" [ 1053.362334] env[61986]: _type = "Task" [ 1053.362334] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.369491] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160361, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.407227] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.407472] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.547161] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746e7d1b-913a-4cc3-b7e3-3773c00ba76f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.554945] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7bf686-f056-405a-8a23-937b8b686820 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.584899] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a55ccd5-1efe-41b7-9679-73be7f4284b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.592054] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c35a12-8deb-43c6-a518-538324f51043 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.604579] env[61986]: DEBUG nova.compute.provider_tree [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.873535] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160361, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056976} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.873826] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.874965] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dabbad-a5c4-4b90-817c-0433e02ebb67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.894956] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.895245] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-689fdbb5-5772-418c-8d3e-0f422be75819 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.914894] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.914894] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1053.915014] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Rebuilding the list of instances to heal {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1053.916595] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1053.916595] env[61986]: value = "task-1160362" [ 1053.916595] env[61986]: _type = "Task" [ 1053.916595] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.924951] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160362, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.107848] env[61986]: DEBUG nova.scheduler.client.report [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.419863] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1054.420202] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Skipping network cache update for instance because it is Building. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1054.433433] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160362, 'name': ReconfigVM_Task, 'duration_secs': 0.424286} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.433808] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Reconfigured VM instance instance-00000060 to attach disk [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.434404] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f2c60fc-a09f-433a-8b7c-4fe79954aea3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.440039] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.440206] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.440355] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1054.440511] env[61986]: DEBUG nova.objects.instance [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lazy-loading 'info_cache' on Instance uuid bc93bdd5-256e-4900-a9ae-609b0a7d402d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.442847] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1054.442847] env[61986]: value = "task-1160363" [ 1054.442847] env[61986]: _type = "Task" [ 1054.442847] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.450576] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160363, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.613257] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.211s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.613812] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.616396] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.308s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.616619] env[61986]: DEBUG nova.objects.instance [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'pci_requests' on Instance uuid 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.955038] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160363, 'name': Rename_Task, 'duration_secs': 0.136533} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.956025] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1054.956143] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62e6c47a-9b10-4002-842e-f8ed378dd34f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.962802] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1054.962802] env[61986]: value = "task-1160364" [ 1054.962802] env[61986]: _type = "Task" [ 1054.962802] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.970251] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160364, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.119635] env[61986]: DEBUG nova.compute.utils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.122442] env[61986]: DEBUG nova.objects.instance [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'numa_topology' on Instance uuid 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.124032] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.124032] env[61986]: DEBUG nova.network.neutron [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1055.170809] env[61986]: DEBUG nova.policy [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c40c9363dc7741a3bd6a040f20284837', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e0f7d6b6be498eaec797aa64aaea7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1055.427440] env[61986]: DEBUG nova.network.neutron [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Successfully created port: b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.467837] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1055.474145] env[61986]: DEBUG oslo_vmware.api [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160364, 'name': PowerOnVM_Task, 'duration_secs': 0.495331} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.474399] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1055.474602] env[61986]: INFO nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Took 4.82 seconds to spawn the instance on the hypervisor. [ 1055.474850] env[61986]: DEBUG nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.475561] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cce76e-cec8-4028-a1d6-2b0f468e186f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.624785] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1055.628864] env[61986]: INFO nova.compute.claims [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.994594] env[61986]: INFO nova.compute.manager [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Took 19.34 seconds to build instance. [ 1056.047068] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.496233] env[61986]: DEBUG oslo_concurrency.lockutils [None req-efbcafdf-e59d-4874-9248-7ecf83913f34 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "fc3ae819-efa0-497e-8d2f-d17a53726056" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.857s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.549998] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-bc93bdd5-256e-4900-a9ae-609b0a7d402d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.550473] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1056.550753] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.550933] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.551093] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.551243] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.551383] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1056.640684] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1056.670308] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.670653] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.670784] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.670972] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.671139] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.671293] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.671521] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.671720] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.671897] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.672075] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.672258] env[61986]: DEBUG nova.virt.hardware [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.673431] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbe5838-b66b-417c-982e-25406af7f54a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.684576] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fd5e56-67ce-4d40-a3fd-9d3ed161627c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.781009] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bf79d3-97ea-4ca4-85dc-286e65d6a1c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.789046] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77ea30f-bf86-45c9-8635-83607c9e7bb2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.820270] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee94c352-8bd9-4dad-8c72-a7e374af4ce5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.827460] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fb829f-2d3c-4205-9539-ef2905a906f8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.840396] env[61986]: DEBUG nova.compute.provider_tree [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.079743] env[61986]: DEBUG nova.compute.manager [req-e52110d8-ebe9-4907-8e0d-b4e3cfb36de9 req-3b6713c8-195b-48f7-81d5-08390b57b831 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Received event network-vif-plugged-b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.079985] env[61986]: DEBUG oslo_concurrency.lockutils [req-e52110d8-ebe9-4907-8e0d-b4e3cfb36de9 req-3b6713c8-195b-48f7-81d5-08390b57b831 service nova] Acquiring lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.080210] env[61986]: DEBUG oslo_concurrency.lockutils [req-e52110d8-ebe9-4907-8e0d-b4e3cfb36de9 req-3b6713c8-195b-48f7-81d5-08390b57b831 service nova] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.080379] env[61986]: DEBUG oslo_concurrency.lockutils [req-e52110d8-ebe9-4907-8e0d-b4e3cfb36de9 req-3b6713c8-195b-48f7-81d5-08390b57b831 service nova] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.080551] env[61986]: DEBUG nova.compute.manager [req-e52110d8-ebe9-4907-8e0d-b4e3cfb36de9 req-3b6713c8-195b-48f7-81d5-08390b57b831 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] No waiting events found dispatching network-vif-plugged-b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.080905] env[61986]: WARNING nova.compute.manager [req-e52110d8-ebe9-4907-8e0d-b4e3cfb36de9 req-3b6713c8-195b-48f7-81d5-08390b57b831 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Received unexpected event network-vif-plugged-b638bc20-c1b8-4047-a795-2d463953e54a for instance with vm_state building and task_state spawning. [ 1057.147174] env[61986]: INFO nova.compute.manager [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Rebuilding instance [ 1057.182827] env[61986]: DEBUG nova.compute.manager [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.183730] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1aefed-4cd1-4427-9424-c4c3c619e28e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.196835] env[61986]: DEBUG nova.network.neutron [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Successfully updated port: b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.343331] env[61986]: DEBUG nova.scheduler.client.report [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.694488] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1057.695019] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-9efa4214-3954-49ef-bc7b-6daba8ba07f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.695157] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-9efa4214-3954-49ef-bc7b-6daba8ba07f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.695298] env[61986]: DEBUG nova.network.neutron [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1057.696381] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8be9f135-f968-44a8-9928-0b9aa2957d6f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.704667] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1057.704667] env[61986]: value = "task-1160365" [ 1057.704667] env[61986]: _type = "Task" [ 1057.704667] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.712998] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160365, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.848395] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.232s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.850520] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.259s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.850742] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.852505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.740s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.852698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.854240] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.308s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.855675] env[61986]: INFO nova.compute.claims [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.880122] env[61986]: INFO nova.scheduler.client.report [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Deleted allocations for instance bc93bdd5-256e-4900-a9ae-609b0a7d402d [ 1057.881823] env[61986]: INFO nova.scheduler.client.report [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Deleted allocations for instance 18893690-92d4-4391-a395-5cf2115397c3 [ 1057.888122] env[61986]: INFO nova.network.neutron [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1058.217204] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160365, 'name': PowerOffVM_Task, 'duration_secs': 0.149841} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.217545] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1058.217810] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1058.218639] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102edec9-693d-4e42-919c-f996b4e5456e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.225135] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1058.225409] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-422ae14d-c3c4-479c-8aa1-19fde0d85198 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.228555] env[61986]: DEBUG nova.network.neutron [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.249888] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1058.250125] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1058.250317] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Deleting the datastore file [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.250578] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9ed3e70-294b-46a4-b152-94b8db703bde {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.256943] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1058.256943] env[61986]: value = "task-1160367" [ 1058.256943] env[61986]: _type = "Task" [ 1058.256943] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.264629] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160367, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.367191] env[61986]: DEBUG nova.network.neutron [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Updating instance_info_cache with network_info: [{"id": "b638bc20-c1b8-4047-a795-2d463953e54a", "address": "fa:16:3e:f0:a0:32", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb638bc20-c1", "ovs_interfaceid": "b638bc20-c1b8-4047-a795-2d463953e54a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.389746] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7cfe17-c242-4955-b18f-7edc0dfe4592 tempest-AttachInterfacesTestJSON-892883867 tempest-AttachInterfacesTestJSON-892883867-project-member] Lock "bc93bdd5-256e-4900-a9ae-609b0a7d402d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.804s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.395957] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b0633418-3bcb-4d7d-9da4-ce00b7b9d0ef tempest-InstanceActionsV221TestJSON-134154326 tempest-InstanceActionsV221TestJSON-134154326-project-member] Lock "18893690-92d4-4391-a395-5cf2115397c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.224s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.767352] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160367, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136673} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.767627] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.767810] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1058.767990] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1058.868050] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-9efa4214-3954-49ef-bc7b-6daba8ba07f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.868401] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Instance network_info: |[{"id": "b638bc20-c1b8-4047-a795-2d463953e54a", "address": "fa:16:3e:f0:a0:32", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb638bc20-c1", "ovs_interfaceid": "b638bc20-c1b8-4047-a795-2d463953e54a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.868637] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:a0:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b638bc20-c1b8-4047-a795-2d463953e54a', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.876170] env[61986]: DEBUG oslo.service.loopingcall [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.876645] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1058.876874] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6bd111f-545c-4de5-939a-c96611327c49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.901557] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.901557] env[61986]: value = "task-1160368" [ 1058.901557] env[61986]: _type = "Task" [ 1058.901557] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.912076] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160368, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.997121] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877d4742-c76b-48f7-83ed-178124fe6c47 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.002902] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237567b7-502d-4672-86a9-e3c34084dc83 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.033828] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4485d837-9986-4ac2-9b64-059540614729 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.041365] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad60ed1f-ce54-4187-ba5d-508e34344eb7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.057704] env[61986]: DEBUG nova.compute.provider_tree [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.113148] env[61986]: DEBUG nova.compute.manager [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Received event network-changed-b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1059.113148] env[61986]: DEBUG nova.compute.manager [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Refreshing instance network info cache due to event network-changed-b638bc20-c1b8-4047-a795-2d463953e54a. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1059.113148] env[61986]: DEBUG oslo_concurrency.lockutils [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] Acquiring lock "refresh_cache-9efa4214-3954-49ef-bc7b-6daba8ba07f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.113148] env[61986]: DEBUG oslo_concurrency.lockutils [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] Acquired lock "refresh_cache-9efa4214-3954-49ef-bc7b-6daba8ba07f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.113148] env[61986]: DEBUG nova.network.neutron [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Refreshing network info cache for port b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.411102] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160368, 'name': CreateVM_Task, 'duration_secs': 0.311936} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.411443] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1059.412284] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.412584] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.412992] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.413352] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a8baec9-96ec-4264-bd4e-90d3d9842d7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.418510] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1059.418510] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5256fb15-d20d-f103-d18a-4370904f75fb" [ 1059.418510] env[61986]: _type = "Task" [ 1059.418510] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.427089] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5256fb15-d20d-f103-d18a-4370904f75fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.563030] env[61986]: DEBUG nova.scheduler.client.report [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.739371] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.739371] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.743023] env[61986]: DEBUG nova.network.neutron [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.811753] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.812094] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.812279] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.812472] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.812628] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.812783] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.812996] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.813183] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.813364] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.813535] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.813738] env[61986]: DEBUG nova.virt.hardware [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.814651] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883d640c-1887-405e-9f41-6d14e7a9c221 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.822969] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a933ff-0e56-48fb-8016-4f78d444620b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.836738] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance VIF info [] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.842471] env[61986]: DEBUG oslo.service.loopingcall [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.842740] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1059.842965] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1961c49-afea-4d2e-aa75-39fcd964a2b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.859438] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.859438] env[61986]: value = "task-1160369" [ 1059.859438] env[61986]: _type = "Task" [ 1059.859438] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.867015] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160369, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.929994] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5256fb15-d20d-f103-d18a-4370904f75fb, 'name': SearchDatastore_Task, 'duration_secs': 0.009268} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.929994] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.929994] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.930816] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.930816] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.930816] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.931781] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f360862-8da2-47dd-8e4e-78bad3c2b972 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.939480] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.939673] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1059.940442] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-317f3ac4-5d6a-416d-84bb-ead09b14affe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.945580] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1059.945580] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527a12a4-420f-43ae-0ba9-8457bae2ac3d" [ 1059.945580] env[61986]: _type = "Task" [ 1059.945580] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.954854] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527a12a4-420f-43ae-0ba9-8457bae2ac3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.023636] env[61986]: DEBUG nova.network.neutron [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Updated VIF entry in instance network info cache for port b638bc20-c1b8-4047-a795-2d463953e54a. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.024021] env[61986]: DEBUG nova.network.neutron [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Updating instance_info_cache with network_info: [{"id": "b638bc20-c1b8-4047-a795-2d463953e54a", "address": "fa:16:3e:f0:a0:32", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb638bc20-c1", "ovs_interfaceid": "b638bc20-c1b8-4047-a795-2d463953e54a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.068940] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.069546] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1060.072288] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.115s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.072504] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.074604] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.912s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.076653] env[61986]: INFO nova.compute.claims [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.128619] env[61986]: INFO nova.scheduler.client.report [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocations for instance eb10ea87-8019-4f3b-af3f-32a901ec6a07 [ 1060.374539] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160369, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.458909] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527a12a4-420f-43ae-0ba9-8457bae2ac3d, 'name': SearchDatastore_Task, 'duration_secs': 0.008963} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.462199] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-216e6f76-450f-4be1-a501-21ed7d3d2d99 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.471424] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1060.471424] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b638b2-0310-31e0-cb58-b764dcd593a5" [ 1060.471424] env[61986]: _type = "Task" [ 1060.471424] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.480941] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b638b2-0310-31e0-cb58-b764dcd593a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.528021] env[61986]: DEBUG oslo_concurrency.lockutils [req-4347be54-9031-479c-a9c4-06263a624699 req-a7c12688-748a-4786-9b3a-9525165ef145 service nova] Releasing lock "refresh_cache-9efa4214-3954-49ef-bc7b-6daba8ba07f2" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.581554] env[61986]: DEBUG nova.compute.utils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1060.586535] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1060.586802] env[61986]: DEBUG nova.network.neutron [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1060.626338] env[61986]: DEBUG nova.policy [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ea19b4c355d4b2797683d78e82eadea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0110dfcd30de444badf6fe685042ba3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1060.635844] env[61986]: DEBUG nova.network.neutron [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.639627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-983d5388-9d22-4333-a867-9a50103a68d9 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "eb10ea87-8019-4f3b-af3f-32a901ec6a07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.622s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.878569] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160369, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.984798] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b638b2-0310-31e0-cb58-b764dcd593a5, 'name': SearchDatastore_Task, 'duration_secs': 0.009985} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.985111] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.985386] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 9efa4214-3954-49ef-bc7b-6daba8ba07f2/9efa4214-3954-49ef-bc7b-6daba8ba07f2.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1060.986078] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b48b240-bfd2-4940-bcaa-3f26d1de48b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.991768] env[61986]: DEBUG nova.network.neutron [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Successfully created port: ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.998276] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1060.998276] env[61986]: value = "task-1160370" [ 1060.998276] env[61986]: _type = "Task" [ 1060.998276] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.007807] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.090021] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1061.141188] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.145363] env[61986]: DEBUG nova.compute.manager [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-vif-plugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.146113] env[61986]: DEBUG oslo_concurrency.lockutils [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.147052] env[61986]: DEBUG oslo_concurrency.lockutils [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.147052] env[61986]: DEBUG oslo_concurrency.lockutils [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.147052] env[61986]: DEBUG nova.compute.manager [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] No waiting events found dispatching network-vif-plugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.147052] env[61986]: WARNING nova.compute.manager [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received unexpected event network-vif-plugged-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 for instance with vm_state shelved_offloaded and task_state spawning. [ 1061.147052] env[61986]: DEBUG nova.compute.manager [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.147052] env[61986]: DEBUG nova.compute.manager [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing instance network info cache due to event network-changed-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1061.147391] env[61986]: DEBUG oslo_concurrency.lockutils [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] Acquiring lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.147391] env[61986]: DEBUG oslo_concurrency.lockutils [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] Acquired lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.148036] env[61986]: DEBUG nova.network.neutron [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Refreshing network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.186249] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='55f0203c1983f9f935f20acef277e9b5',container_format='bare',created_at=2024-10-07T06:45:57Z,direct_url=,disk_format='vmdk',id=15c86942-389c-4a86-952d-179d0c1d0572,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-493125558-shelved',owner='691d4935610e45e6a0bae5ca5f1197f3',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-07T06:46:10Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.187675] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.188364] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.188994] env[61986]: DEBUG nova.virt.hardware [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.189996] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c50285e-f436-45b3-acd8-f2563ea68ce0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.201934] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4b2d4f-48da-4d0d-825e-3f9bcc80dc02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.223231] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:9b:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1768af3d-3317-4ef5-b484-0c2707d63de7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1061.231129] env[61986]: DEBUG oslo.service.loopingcall [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.235237] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1061.236332] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-461a9ba4-8307-4361-a0d4-fb894df882fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.252389] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0188fa79-7d43-4647-911a-5f3fff0dd11c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.264483] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101db455-6b78-4e0a-b491-aabb01ac1ec0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.264959] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1061.264959] env[61986]: value = "task-1160371" [ 1061.264959] env[61986]: _type = "Task" [ 1061.264959] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.295963] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7eaaf8d-be38-4c5c-9b95-a16b2d143cdf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.302839] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160371, 'name': CreateVM_Task} progress is 15%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.308035] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379746e4-ac41-4aea-aec4-cf627001ba68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.323021] env[61986]: DEBUG nova.compute.provider_tree [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.373676] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160369, 'name': CreateVM_Task, 'duration_secs': 1.252838} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.373764] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1061.374827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.374827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.374827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1061.375062] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6173e40e-cf6e-4dad-b91b-4dbddce04544 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.380281] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1061.380281] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522bf375-49de-e224-14fc-7c952f40e588" [ 1061.380281] env[61986]: _type = "Task" [ 1061.380281] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.390728] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522bf375-49de-e224-14fc-7c952f40e588, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.511639] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490787} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.511996] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 9efa4214-3954-49ef-bc7b-6daba8ba07f2/9efa4214-3954-49ef-bc7b-6daba8ba07f2.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1061.512223] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.512554] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c02ef321-fedd-43c6-94ca-6faf5049d5b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.518781] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1061.518781] env[61986]: value = "task-1160372" [ 1061.518781] env[61986]: _type = "Task" [ 1061.518781] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.528637] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160372, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.776010] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160371, 'name': CreateVM_Task, 'duration_secs': 0.338004} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.776066] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1061.776701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.776893] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.777254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1061.779614] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-879985c4-58c2-40b6-ab52-4c86b263beda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.784285] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1061.784285] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228c864-6624-7dc5-898b-c0d3fc9ee73a" [ 1061.784285] env[61986]: _type = "Task" [ 1061.784285] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.792672] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5228c864-6624-7dc5-898b-c0d3fc9ee73a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.826262] env[61986]: DEBUG nova.scheduler.client.report [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.875215] env[61986]: DEBUG nova.network.neutron [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updated VIF entry in instance network info cache for port a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.875629] env[61986]: DEBUG nova.network.neutron [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [{"id": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "address": "fa:16:3e:a2:9b:bc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa298ab4e-6b", "ovs_interfaceid": "a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.889372] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522bf375-49de-e224-14fc-7c952f40e588, 'name': SearchDatastore_Task, 'duration_secs': 0.053137} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.889670] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.889873] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.890117] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.890269] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.890447] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.890741] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dea0ea3f-7693-4886-9dc8-83b5d214594c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.906911] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.907116] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1061.907820] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2d0b571-4d45-4834-923a-b6b751ae5cdf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.912682] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1061.912682] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a3a245-80ee-9da9-027c-7e6131a6bb99" [ 1061.912682] env[61986]: _type = "Task" [ 1061.912682] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.920259] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a3a245-80ee-9da9-027c-7e6131a6bb99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.028877] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160372, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065739} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.029228] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.030015] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64234f0e-260a-4c22-8bde-84423b993b59 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.051794] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 9efa4214-3954-49ef-bc7b-6daba8ba07f2/9efa4214-3954-49ef-bc7b-6daba8ba07f2.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.052074] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07c8ea17-498d-498a-aefe-afd702419c68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.071379] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1062.071379] env[61986]: value = "task-1160373" [ 1062.071379] env[61986]: _type = "Task" [ 1062.071379] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.083355] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160373, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.099865] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.130370] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.130665] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.130881] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.131113] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.131291] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.131468] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.131710] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.131914] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.132121] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.132322] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.132528] env[61986]: DEBUG nova.virt.hardware [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.133464] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a60e96d-d61e-4916-b412-35102076ab2d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.143644] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92367220-ad9a-456b-81c4-5b5ea01e3a61 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.295177] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.295567] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Processing image 15c86942-389c-4a86-952d-179d0c1d0572 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1062.295725] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.295895] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.296105] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1062.296370] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5015372-6175-48aa-b195-02ad12709e8d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.304761] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.305057] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1062.305682] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4229897d-bd99-4722-8fed-dc38583a285d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.310874] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1062.310874] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521702d1-dfc6-ff86-1ceb-e91c7473e205" [ 1062.310874] env[61986]: _type = "Task" [ 1062.310874] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.324328] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1062.324579] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Fetch image to [datastore2] OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be/OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1062.324766] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Downloading stream optimized image 15c86942-389c-4a86-952d-179d0c1d0572 to [datastore2] OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be/OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be.vmdk on the data store datastore2 as vApp {{(pid=61986) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1062.324940] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Downloading image file data 15c86942-389c-4a86-952d-179d0c1d0572 to the ESX as VM named 'OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be' {{(pid=61986) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1062.330548] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.331086] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1062.378158] env[61986]: DEBUG oslo_concurrency.lockutils [req-c2c2beb4-5abb-4206-a5a2-595abaf6116e req-650b5cd4-ae30-422c-bd5f-65a46e45c4d8 service nova] Releasing lock "refresh_cache-221e10a3-da31-410c-80f8-4bcc2c515710" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.423754] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a3a245-80ee-9da9-027c-7e6131a6bb99, 'name': SearchDatastore_Task, 'duration_secs': 0.007795} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.424583] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc2e8ae7-5cdf-4064-aaa6-1c609944c6e0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.427533] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1062.427533] env[61986]: value = "resgroup-9" [ 1062.427533] env[61986]: _type = "ResourcePool" [ 1062.427533] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1062.427775] env[61986]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5449f3c9-072b-4064-8e05-a59884e440fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.446895] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1062.446895] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524cd021-dbac-afc4-888d-25bbbba05f06" [ 1062.446895] env[61986]: _type = "Task" [ 1062.446895] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.451203] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease: (returnval){ [ 1062.451203] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521105de-7e76-ccff-14c7-7b20e8ca910d" [ 1062.451203] env[61986]: _type = "HttpNfcLease" [ 1062.451203] env[61986]: } obtained for vApp import into resource pool (val){ [ 1062.451203] env[61986]: value = "resgroup-9" [ 1062.451203] env[61986]: _type = "ResourcePool" [ 1062.451203] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1062.451435] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the lease: (returnval){ [ 1062.451435] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521105de-7e76-ccff-14c7-7b20e8ca910d" [ 1062.451435] env[61986]: _type = "HttpNfcLease" [ 1062.451435] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1062.457707] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524cd021-dbac-afc4-888d-25bbbba05f06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.460766] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1062.460766] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521105de-7e76-ccff-14c7-7b20e8ca910d" [ 1062.460766] env[61986]: _type = "HttpNfcLease" [ 1062.460766] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1062.581984] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160373, 'name': ReconfigVM_Task, 'duration_secs': 0.265479} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.582300] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 9efa4214-3954-49ef-bc7b-6daba8ba07f2/9efa4214-3954-49ef-bc7b-6daba8ba07f2.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.582961] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f6181b8-26b5-4ad2-885f-8c35cb8cec7b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.589082] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1062.589082] env[61986]: value = "task-1160375" [ 1062.589082] env[61986]: _type = "Task" [ 1062.589082] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.600613] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160375, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.836184] env[61986]: DEBUG nova.compute.utils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.837965] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1062.838260] env[61986]: DEBUG nova.network.neutron [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1062.878513] env[61986]: DEBUG nova.policy [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3067ff88a6d94daea3e7f5488284457b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5342ea3a8f664440a608b16f1b7b44ba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1062.960619] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524cd021-dbac-afc4-888d-25bbbba05f06, 'name': SearchDatastore_Task, 'duration_secs': 0.017946} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.961469] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.961741] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1062.961990] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84f64e32-31c1-4195-a8da-69923392b196 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.965252] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1062.965252] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521105de-7e76-ccff-14c7-7b20e8ca910d" [ 1062.965252] env[61986]: _type = "HttpNfcLease" [ 1062.965252] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1062.965815] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1062.965815] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521105de-7e76-ccff-14c7-7b20e8ca910d" [ 1062.965815] env[61986]: _type = "HttpNfcLease" [ 1062.965815] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1062.966711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0030ec61-d560-47f9-9924-8bc806f857de {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.971044] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1062.971044] env[61986]: value = "task-1160376" [ 1062.971044] env[61986]: _type = "Task" [ 1062.971044] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.976471] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524db48a-735a-1ca2-a6ed-faccc2be8616/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1062.976664] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524db48a-735a-1ca2-a6ed-faccc2be8616/disk-0.vmdk. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1063.041023] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160376, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.045821] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7615c17c-197c-4d77-981e-8ade184743db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.098393] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160375, 'name': Rename_Task, 'duration_secs': 0.140079} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.098717] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1063.098973] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04042829-3bba-4f4e-875c-c3c0fd17fffb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.105922] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1063.105922] env[61986]: value = "task-1160377" [ 1063.105922] env[61986]: _type = "Task" [ 1063.105922] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.113662] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160377, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.153512] env[61986]: DEBUG nova.network.neutron [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Successfully created port: 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.166817] env[61986]: DEBUG nova.network.neutron [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Successfully updated port: ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.202649] env[61986]: DEBUG nova.compute.manager [req-45c7dd56-3ab4-4bc1-a793-e66a10dd4475 req-45ef6036-9c8e-41c2-85d4-8664bd532403 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Received event network-vif-plugged-ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.203298] env[61986]: DEBUG oslo_concurrency.lockutils [req-45c7dd56-3ab4-4bc1-a793-e66a10dd4475 req-45ef6036-9c8e-41c2-85d4-8664bd532403 service nova] Acquiring lock "c3bcd1ba-a3df-4b81-9743-431235536642-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.204218] env[61986]: DEBUG oslo_concurrency.lockutils [req-45c7dd56-3ab4-4bc1-a793-e66a10dd4475 req-45ef6036-9c8e-41c2-85d4-8664bd532403 service nova] Lock "c3bcd1ba-a3df-4b81-9743-431235536642-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.204444] env[61986]: DEBUG oslo_concurrency.lockutils [req-45c7dd56-3ab4-4bc1-a793-e66a10dd4475 req-45ef6036-9c8e-41c2-85d4-8664bd532403 service nova] Lock "c3bcd1ba-a3df-4b81-9743-431235536642-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.204637] env[61986]: DEBUG nova.compute.manager [req-45c7dd56-3ab4-4bc1-a793-e66a10dd4475 req-45ef6036-9c8e-41c2-85d4-8664bd532403 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] No waiting events found dispatching network-vif-plugged-ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.204815] env[61986]: WARNING nova.compute.manager [req-45c7dd56-3ab4-4bc1-a793-e66a10dd4475 req-45ef6036-9c8e-41c2-85d4-8664bd532403 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Received unexpected event network-vif-plugged-ae574939-8249-4b31-b7b3-239d0d191f4b for instance with vm_state building and task_state spawning. [ 1063.341356] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1063.491050] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160376, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516784} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.491420] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1063.491722] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.492178] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40bac344-bf45-4ae7-a88f-29da5789ab95 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.501089] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1063.501089] env[61986]: value = "task-1160378" [ 1063.501089] env[61986]: _type = "Task" [ 1063.501089] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.513030] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160378, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.619719] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160377, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.670407] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "refresh_cache-c3bcd1ba-a3df-4b81-9743-431235536642" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.670407] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "refresh_cache-c3bcd1ba-a3df-4b81-9743-431235536642" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.670512] env[61986]: DEBUG nova.network.neutron [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1064.011428] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160378, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124146} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.012837] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.012837] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a82bb5b-97db-44e3-90d1-1b1e82dd2625 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.034793] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.039706] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3619acc0-fa5d-4c35-b591-f18e7fafdc7f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.064241] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1064.064241] env[61986]: value = "task-1160379" [ 1064.064241] env[61986]: _type = "Task" [ 1064.064241] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.076738] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160379, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.120405] env[61986]: DEBUG oslo_vmware.api [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160377, 'name': PowerOnVM_Task, 'duration_secs': 0.634408} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.122164] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1064.122415] env[61986]: INFO nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1064.122605] env[61986]: DEBUG nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.123529] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae45d3f-0ef0-4785-bddd-dc751549d033 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.225101] env[61986]: DEBUG nova.network.neutron [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1064.354661] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1064.378699] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.378993] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.379176] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.379362] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.379514] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.379667] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.379885] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.380193] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.380415] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.380592] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.380858] env[61986]: DEBUG nova.virt.hardware [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.381869] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7067c88-9fdb-4389-92a9-193595bb4092 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.390814] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c0259c-4fb0-4cc8-899c-faccf3d3fa97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.529024] env[61986]: DEBUG nova.network.neutron [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Updating instance_info_cache with network_info: [{"id": "ae574939-8249-4b31-b7b3-239d0d191f4b", "address": "fa:16:3e:28:fe:18", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae574939-82", "ovs_interfaceid": "ae574939-8249-4b31-b7b3-239d0d191f4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.536951] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1064.537552] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524db48a-735a-1ca2-a6ed-faccc2be8616/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1064.538601] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dac443-f5ea-4914-96b5-bfdbf6dd92a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.546266] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524db48a-735a-1ca2-a6ed-faccc2be8616/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1064.546520] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524db48a-735a-1ca2-a6ed-faccc2be8616/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1064.547526] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b4fe55de-fae5-4eee-a71b-089ac4afaa42 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.577628] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160379, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.646267] env[61986]: INFO nova.compute.manager [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Took 24.21 seconds to build instance. [ 1064.713464] env[61986]: DEBUG nova.compute.manager [req-92325385-eb4b-4b8a-8a40-9ccb2293fae6 req-f6d2bb36-33d7-4549-aab1-ab404b7c5d0d service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-vif-plugged-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1064.714009] env[61986]: DEBUG oslo_concurrency.lockutils [req-92325385-eb4b-4b8a-8a40-9ccb2293fae6 req-f6d2bb36-33d7-4549-aab1-ab404b7c5d0d service nova] Acquiring lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.718214] env[61986]: DEBUG oslo_concurrency.lockutils [req-92325385-eb4b-4b8a-8a40-9ccb2293fae6 req-f6d2bb36-33d7-4549-aab1-ab404b7c5d0d service nova] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.718214] env[61986]: DEBUG oslo_concurrency.lockutils [req-92325385-eb4b-4b8a-8a40-9ccb2293fae6 req-f6d2bb36-33d7-4549-aab1-ab404b7c5d0d service nova] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.718214] env[61986]: DEBUG nova.compute.manager [req-92325385-eb4b-4b8a-8a40-9ccb2293fae6 req-f6d2bb36-33d7-4549-aab1-ab404b7c5d0d service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] No waiting events found dispatching network-vif-plugged-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1064.718214] env[61986]: WARNING nova.compute.manager [req-92325385-eb4b-4b8a-8a40-9ccb2293fae6 req-f6d2bb36-33d7-4549-aab1-ab404b7c5d0d service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received unexpected event network-vif-plugged-12d8dd0f-1d73-4e30-bb9e-f4e074639096 for instance with vm_state building and task_state spawning. [ 1064.768816] env[61986]: DEBUG oslo_vmware.rw_handles [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524db48a-735a-1ca2-a6ed-faccc2be8616/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1064.769229] env[61986]: INFO nova.virt.vmwareapi.images [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Downloaded image file data 15c86942-389c-4a86-952d-179d0c1d0572 [ 1064.770760] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29de37f5-d617-4cf3-b930-5033481d2eef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.805454] env[61986]: DEBUG nova.network.neutron [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Successfully updated port: 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.807981] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9176a9fb-cc43-4a00-adb2-b8b56ae8e0b2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.841602] env[61986]: INFO nova.virt.vmwareapi.images [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] The imported VM was unregistered [ 1064.841602] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1064.841602] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating directory with path [datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.841602] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c96cad17-b7cc-425d-b207-ee32c3676e88 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.862258] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created directory with path [datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.862481] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be/OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be.vmdk to [datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk. {{(pid=61986) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1064.862756] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-027ab926-2fef-4c42-b2e7-83fa0aef1ee9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.871051] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1064.871051] env[61986]: value = "task-1160381" [ 1064.871051] env[61986]: _type = "Task" [ 1064.871051] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.883035] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160381, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.031382] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "refresh_cache-c3bcd1ba-a3df-4b81-9743-431235536642" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.031814] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Instance network_info: |[{"id": "ae574939-8249-4b31-b7b3-239d0d191f4b", "address": "fa:16:3e:28:fe:18", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae574939-82", "ovs_interfaceid": "ae574939-8249-4b31-b7b3-239d0d191f4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1065.032361] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:fe:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae574939-8249-4b31-b7b3-239d0d191f4b', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.041319] env[61986]: DEBUG oslo.service.loopingcall [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.041319] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1065.041560] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd306868-b469-44c8-95a2-dbf10a621467 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.067549] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.067549] env[61986]: value = "task-1160382" [ 1065.067549] env[61986]: _type = "Task" [ 1065.067549] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.082706] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160382, 'name': CreateVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.089262] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160379, 'name': ReconfigVM_Task, 'duration_secs': 0.610848} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.089262] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Reconfigured VM instance instance-00000060 to attach disk [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056/fc3ae819-efa0-497e-8d2f-d17a53726056.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.089262] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92cab001-0d79-4167-9b26-a93f0c915670 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.096175] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1065.096175] env[61986]: value = "task-1160383" [ 1065.096175] env[61986]: _type = "Task" [ 1065.096175] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.108621] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160383, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.148562] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e6809806-37cc-4a54-8ea2-5ecf5e9ec72b tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.721s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.239553] env[61986]: DEBUG nova.compute.manager [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Received event network-changed-ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1065.239553] env[61986]: DEBUG nova.compute.manager [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Refreshing instance network info cache due to event network-changed-ae574939-8249-4b31-b7b3-239d0d191f4b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1065.239736] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] Acquiring lock "refresh_cache-c3bcd1ba-a3df-4b81-9743-431235536642" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.239918] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] Acquired lock "refresh_cache-c3bcd1ba-a3df-4b81-9743-431235536642" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.240164] env[61986]: DEBUG nova.network.neutron [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Refreshing network info cache for port ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1065.308426] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.308657] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.308753] env[61986]: DEBUG nova.network.neutron [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1065.316311] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.317115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.383367] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160381, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.579961] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160382, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.608466] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160383, 'name': Rename_Task, 'duration_secs': 0.222393} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.608777] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1065.610695] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22aeddb0-d3b1-465f-bc97-e43f38670821 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.622031] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1065.622031] env[61986]: value = "task-1160384" [ 1065.622031] env[61986]: _type = "Task" [ 1065.622031] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.633935] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160384, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.819936] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1065.850830] env[61986]: DEBUG nova.network.neutron [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1065.886180] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160381, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.996936] env[61986]: DEBUG nova.network.neutron [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Updated VIF entry in instance network info cache for port ae574939-8249-4b31-b7b3-239d0d191f4b. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1065.997338] env[61986]: DEBUG nova.network.neutron [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Updating instance_info_cache with network_info: [{"id": "ae574939-8249-4b31-b7b3-239d0d191f4b", "address": "fa:16:3e:28:fe:18", "network": {"id": "ca264996-e633-44f3-bdf4-3ce9a719a54d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1477244074-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0110dfcd30de444badf6fe685042ba3b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae574939-82", "ovs_interfaceid": "ae574939-8249-4b31-b7b3-239d0d191f4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.073272] env[61986]: DEBUG nova.network.neutron [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.081176] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160382, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.136641] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160384, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.154596] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.154845] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.347459] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.347779] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.349681] env[61986]: INFO nova.compute.claims [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.385837] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160381, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.501021] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ca18fd2-cce3-47d5-9dde-cde90df64199 req-77a24a91-d58f-45d6-9dbe-787b3568eb29 service nova] Releasing lock "refresh_cache-c3bcd1ba-a3df-4b81-9743-431235536642" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.576082] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.576425] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Instance network_info: |[{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1066.576801] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:a5:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12d8dd0f-1d73-4e30-bb9e-f4e074639096', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.584680] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Creating folder: Project (5342ea3a8f664440a608b16f1b7b44ba). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1066.589160] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5561fd17-e862-4409-8f82-6e163600f926 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.591250] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160382, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.604055] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Created folder: Project (5342ea3a8f664440a608b16f1b7b44ba) in parent group-v252271. [ 1066.604339] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Creating folder: Instances. Parent ref: group-v252460. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1066.604675] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-692c26f4-e3a8-4698-8df9-b300926ed23a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.619694] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Created folder: Instances in parent group-v252460. [ 1066.620079] env[61986]: DEBUG oslo.service.loopingcall [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.620331] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1066.620571] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-587d8800-fd95-40c5-bc58-a4704610f6d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.645465] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160384, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.651556] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.651556] env[61986]: value = "task-1160387" [ 1066.651556] env[61986]: _type = "Task" [ 1066.651556] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.658486] env[61986]: DEBUG nova.compute.utils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1066.663186] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160387, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.863994] env[61986]: DEBUG nova.compute.manager [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1066.864420] env[61986]: DEBUG nova.compute.manager [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing instance network info cache due to event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1066.864546] env[61986]: DEBUG oslo_concurrency.lockutils [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.864726] env[61986]: DEBUG oslo_concurrency.lockutils [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.864966] env[61986]: DEBUG nova.network.neutron [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1066.888084] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160381, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.081510] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160382, 'name': CreateVM_Task, 'duration_secs': 2.009376} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.081908] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1067.082755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.082755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.082956] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.083212] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d16950e-aef5-4cf0-a1fb-4ed7fd54af2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.091623] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1067.091623] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52af41be-749d-ad4b-01cd-c11bbfaba96f" [ 1067.091623] env[61986]: _type = "Task" [ 1067.091623] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.104923] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52af41be-749d-ad4b-01cd-c11bbfaba96f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.152739] env[61986]: DEBUG oslo_vmware.api [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160384, 'name': PowerOnVM_Task, 'duration_secs': 1.047749} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.156332] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.156582] env[61986]: DEBUG nova.compute.manager [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.157497] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdea1c3-7f7d-4395-b4d6-99b66125db88 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.166189] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.175096] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160387, 'name': CreateVM_Task, 'duration_secs': 0.429399} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.176426] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1067.177631] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.384550] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160381, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.358963} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.386956] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be/OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be.vmdk to [datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk. [ 1067.387177] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Cleaning up location [datastore2] OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1067.387348] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a5430b49-4806-449d-afb2-726ca5e248be {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.388420] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e82972ca-2f6b-4a13-a440-ee9b284c9ea0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.394785] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1067.394785] env[61986]: value = "task-1160388" [ 1067.394785] env[61986]: _type = "Task" [ 1067.394785] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.406815] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.467883] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9b81eb-f57a-4b41-9ab6-c6d9569af073 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.478064] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef37a5c0-b32e-406f-92b0-924c33235988 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.509272] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677fc59d-4357-4c71-9cc3-1861180b33d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.517069] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb1ed68-f154-4e68-987e-9bbe593c8fbb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.530279] env[61986]: DEBUG nova.compute.provider_tree [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.572879] env[61986]: DEBUG nova.network.neutron [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updated VIF entry in instance network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1067.573318] env[61986]: DEBUG nova.network.neutron [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.603239] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52af41be-749d-ad4b-01cd-c11bbfaba96f, 'name': SearchDatastore_Task, 'duration_secs': 0.082516} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.603510] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.603747] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.603982] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.604147] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.604331] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.604605] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.604907] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.605145] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea5361e1-8738-4266-bebc-c146aa3a2e72 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.606976] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d999877f-2ff7-475e-9fde-20cdafbea0f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.612177] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1067.612177] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524c1816-8a9a-2240-70eb-dcb280a7ec22" [ 1067.612177] env[61986]: _type = "Task" [ 1067.612177] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.616245] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.616425] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1067.619725] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-187e5d5e-fdfd-4626-acdb-b48902e3654f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.622273] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524c1816-8a9a-2240-70eb-dcb280a7ec22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.625220] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1067.625220] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cad317-a833-a9ae-bb4b-f9978863ab49" [ 1067.625220] env[61986]: _type = "Task" [ 1067.625220] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.632698] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cad317-a833-a9ae-bb4b-f9978863ab49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.684885] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.905637] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160388, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033333} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.905805] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.905980] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.906246] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk to [datastore2] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1067.906502] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26e5ba9f-1ef8-4e7a-97c3-4bfd7c3468a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.913334] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1067.913334] env[61986]: value = "task-1160389" [ 1067.913334] env[61986]: _type = "Task" [ 1067.913334] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.921545] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160389, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.033637] env[61986]: DEBUG nova.scheduler.client.report [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.075391] env[61986]: DEBUG oslo_concurrency.lockutils [req-ae80d2ed-721e-4528-920e-b838a0f03d8e req-52d29d14-ba05-4891-b90f-da0fe0bfe12e service nova] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.123667] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]524c1816-8a9a-2240-70eb-dcb280a7ec22, 'name': SearchDatastore_Task, 'duration_secs': 0.009511} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.124009] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.124293] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.124489] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.135572] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52cad317-a833-a9ae-bb4b-f9978863ab49, 'name': SearchDatastore_Task, 'duration_secs': 0.008456} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.136671] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30305ccd-c0c6-4162-959e-50a4d2503923 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.142318] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1068.142318] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e619d9-1d32-ec79-b4bf-329d4a3220c9" [ 1068.142318] env[61986]: _type = "Task" [ 1068.142318] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.151190] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e619d9-1d32-ec79-b4bf-329d4a3220c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.287023] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.287145] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.287352] env[61986]: INFO nova.compute.manager [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Attaching volume 68570cec-bfcf-4107-82c7-cbbf826dca77 to /dev/sdb [ 1068.323235] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5846d656-4e32-4e9e-9af8-6861c161e1c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.331595] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a250d61e-4f49-4216-9c24-663457632fc0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.349687] env[61986]: DEBUG nova.virt.block_device [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Updating existing volume attachment record: 8fca1641-46aa-4016-afa8-e5977efb0430 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1068.424317] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160389, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.539810] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.192s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.540544] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1068.543761] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.859s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.543969] env[61986]: DEBUG nova.objects.instance [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1068.567185] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "fc3ae819-efa0-497e-8d2f-d17a53726056" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.567480] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "fc3ae819-efa0-497e-8d2f-d17a53726056" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.567698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "fc3ae819-efa0-497e-8d2f-d17a53726056-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.567887] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "fc3ae819-efa0-497e-8d2f-d17a53726056-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.568075] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "fc3ae819-efa0-497e-8d2f-d17a53726056-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.570817] env[61986]: INFO nova.compute.manager [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Terminating instance [ 1068.573108] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "refresh_cache-fc3ae819-efa0-497e-8d2f-d17a53726056" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.573306] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquired lock "refresh_cache-fc3ae819-efa0-497e-8d2f-d17a53726056" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.573551] env[61986]: DEBUG nova.network.neutron [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.655076] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e619d9-1d32-ec79-b4bf-329d4a3220c9, 'name': SearchDatastore_Task, 'duration_secs': 0.049045} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.655076] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.655540] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] c3bcd1ba-a3df-4b81-9743-431235536642/c3bcd1ba-a3df-4b81-9743-431235536642.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1068.655996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.656470] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.656934] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aabee00a-6976-4e35-859e-1211828be4ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.659583] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e746345c-b10d-42af-90a9-267f7462d25f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.669068] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1068.669068] env[61986]: value = "task-1160393" [ 1068.669068] env[61986]: _type = "Task" [ 1068.669068] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.678045] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.683979] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.684215] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1068.685071] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-343c9189-9512-4fb7-9967-5043f294b95d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.693717] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1068.693717] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52dbd674-4753-1072-5146-298e77c2a114" [ 1068.693717] env[61986]: _type = "Task" [ 1068.693717] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.705273] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52dbd674-4753-1072-5146-298e77c2a114, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.927430] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160389, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.049433] env[61986]: DEBUG nova.compute.utils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1069.056233] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1069.057282] env[61986]: DEBUG nova.network.neutron [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1069.094621] env[61986]: DEBUG nova.network.neutron [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1069.136319] env[61986]: DEBUG nova.policy [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1069.156694] env[61986]: DEBUG nova.network.neutron [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.184339] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.209549] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52dbd674-4753-1072-5146-298e77c2a114, 'name': SearchDatastore_Task, 'duration_secs': 0.098208} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.210441] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c346fc0-9e0e-48ff-9db5-64e14bda6dac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.218206] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1069.218206] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527c59c9-db3f-5d79-aba3-72658dcc783d" [ 1069.218206] env[61986]: _type = "Task" [ 1069.218206] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.228973] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527c59c9-db3f-5d79-aba3-72658dcc783d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.429020] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160389, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.473857] env[61986]: DEBUG nova.network.neutron [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Successfully created port: 73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.558049] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1069.562473] env[61986]: DEBUG oslo_concurrency.lockutils [None req-74bc59f7-4632-4f9d-90b6-21f9e3589b56 tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.662410] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Releasing lock "refresh_cache-fc3ae819-efa0-497e-8d2f-d17a53726056" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.662410] env[61986]: DEBUG nova.compute.manager [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.662410] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1069.663136] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8539ca-a875-490d-8c4b-9564b47838fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.678091] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1069.678892] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f134887-b8b6-47d8-b5e6-a9cd29174805 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.684564] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.689057] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1069.689057] env[61986]: value = "task-1160394" [ 1069.689057] env[61986]: _type = "Task" [ 1069.689057] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.701424] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160394, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.731895] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527c59c9-db3f-5d79-aba3-72658dcc783d, 'name': SearchDatastore_Task, 'duration_secs': 0.086602} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.732284] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.732555] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/18f87ccd-ca34-44f6-aa7d-d38397cc479b.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1069.732838] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b901f61-05d7-424c-9ce5-c3e4cd6a4ad0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.743756] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1069.743756] env[61986]: value = "task-1160395" [ 1069.743756] env[61986]: _type = "Task" [ 1069.743756] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.930039] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160389, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.184303] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160393, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.204635] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160394, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.255169] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160395, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.429757] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160389, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.356591} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.430132] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/15c86942-389c-4a86-952d-179d0c1d0572/15c86942-389c-4a86-952d-179d0c1d0572.vmdk to [datastore2] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1070.431118] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec93c103-c3d6-4ede-b46f-8db6c8656d30 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.456388] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.456778] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f75e464-1c95-4447-ae33-d5b36e6a2cc7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.479549] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1070.479549] env[61986]: value = "task-1160396" [ 1070.479549] env[61986]: _type = "Task" [ 1070.479549] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.491099] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.570924] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1070.598869] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.599233] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.599437] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.599713] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.599913] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.600148] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.600394] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.600583] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.600869] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.601120] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.601375] env[61986]: DEBUG nova.virt.hardware [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.602362] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71074cfe-cef8-4500-a1d5-c4665171b41d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.613558] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e1aa72-a71d-47e2-b16e-586c2f287ad4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.683257] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160393, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.898892} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.683919] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] c3bcd1ba-a3df-4b81-9743-431235536642/c3bcd1ba-a3df-4b81-9743-431235536642.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1070.684289] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1070.684678] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ec48d64-ece2-4041-b7bf-a2591b8396db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.697066] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1070.697066] env[61986]: value = "task-1160398" [ 1070.697066] env[61986]: _type = "Task" [ 1070.697066] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.705301] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160394, 'name': PowerOffVM_Task, 'duration_secs': 0.889114} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.706205] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1070.706416] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1070.706776] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b0e4a40-7aac-479b-8131-c9150697bc8a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.712837] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160398, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.739260] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1070.739260] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1070.739260] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Deleting the datastore file [datastore2] fc3ae819-efa0-497e-8d2f-d17a53726056 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.739260] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7900bdc3-a7cb-400f-bf70-1527ed8ece1e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.746859] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for the task: (returnval){ [ 1070.746859] env[61986]: value = "task-1160400" [ 1070.746859] env[61986]: _type = "Task" [ 1070.746859] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.760447] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160395, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.764035] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.989413] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160396, 'name': ReconfigVM_Task, 'duration_secs': 0.352583} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.989704] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 221e10a3-da31-410c-80f8-4bcc2c515710/221e10a3-da31-410c-80f8-4bcc2c515710.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.991324] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encrypted': False, 'size': 0, 'encryption_secret_uuid': None, 'encryption_format': None, 'boot_index': 0, 'disk_bus': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'encryption_options': None, 'image_id': '7b73cf0e-555e-4258-a7ba-6cecbc90b6c5'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252451', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'name': 'volume-3f7d5322-8141-4aa7-813d-2ae4f033983e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '221e10a3-da31-410c-80f8-4bcc2c515710', 'attached_at': '', 'detached_at': '', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'serial': '3f7d5322-8141-4aa7-813d-2ae4f033983e'}, 'guest_format': None, 'attachment_id': '136a5beb-c6fc-4d2e-bb53-de40e7b5f254', 'boot_index': None, 'disk_bus': None, 'mount_device': '/dev/sdb', 'device_type': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61986) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1070.991324] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1070.991532] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252451', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'name': 'volume-3f7d5322-8141-4aa7-813d-2ae4f033983e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '221e10a3-da31-410c-80f8-4bcc2c515710', 'attached_at': '', 'detached_at': '', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'serial': '3f7d5322-8141-4aa7-813d-2ae4f033983e'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1070.992292] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b61d7a3-16a6-42e7-b904-177555355664 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.997252] env[61986]: DEBUG nova.compute.manager [req-49a51e21-6b0b-4e88-90be-b10facbd0a77 req-441f2c4c-8e40-43e0-a591-a4ec4d252f45 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Received event network-vif-plugged-73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1070.997457] env[61986]: DEBUG oslo_concurrency.lockutils [req-49a51e21-6b0b-4e88-90be-b10facbd0a77 req-441f2c4c-8e40-43e0-a591-a4ec4d252f45 service nova] Acquiring lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.997662] env[61986]: DEBUG oslo_concurrency.lockutils [req-49a51e21-6b0b-4e88-90be-b10facbd0a77 req-441f2c4c-8e40-43e0-a591-a4ec4d252f45 service nova] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.997831] env[61986]: DEBUG oslo_concurrency.lockutils [req-49a51e21-6b0b-4e88-90be-b10facbd0a77 req-441f2c4c-8e40-43e0-a591-a4ec4d252f45 service nova] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.998007] env[61986]: DEBUG nova.compute.manager [req-49a51e21-6b0b-4e88-90be-b10facbd0a77 req-441f2c4c-8e40-43e0-a591-a4ec4d252f45 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] No waiting events found dispatching network-vif-plugged-73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.998185] env[61986]: WARNING nova.compute.manager [req-49a51e21-6b0b-4e88-90be-b10facbd0a77 req-441f2c4c-8e40-43e0-a591-a4ec4d252f45 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Received unexpected event network-vif-plugged-73c2047a-043c-4e3a-ae94-b9749f2229f3 for instance with vm_state building and task_state spawning. [ 1071.012088] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35755721-4c8a-49da-88bd-1e39c82d54f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.036847] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] volume-3f7d5322-8141-4aa7-813d-2ae4f033983e/volume-3f7d5322-8141-4aa7-813d-2ae4f033983e.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.037235] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8cb0f75-d87e-4540-9045-38d028473a5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.056802] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1071.056802] env[61986]: value = "task-1160401" [ 1071.056802] env[61986]: _type = "Task" [ 1071.056802] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.065453] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160401, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.079084] env[61986]: DEBUG nova.network.neutron [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Successfully updated port: 73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1071.208242] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160398, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.167963} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.208585] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.209553] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d006e4b2-eb2e-4f19-9dab-43e525fe1e97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.237296] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] c3bcd1ba-a3df-4b81-9743-431235536642/c3bcd1ba-a3df-4b81-9743-431235536642.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.237626] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e20e8b3f-7622-4320-b9ab-b4d8093a358a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.264550] env[61986]: DEBUG oslo_vmware.api [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Task: {'id': task-1160400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222807} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.268355] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.268570] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1071.268884] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.269008] env[61986]: INFO nova.compute.manager [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1071.269264] env[61986]: DEBUG oslo.service.loopingcall [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.269523] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1071.269523] env[61986]: value = "task-1160402" [ 1071.269523] env[61986]: _type = "Task" [ 1071.269523] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.269723] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160395, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.158788} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.269925] env[61986]: DEBUG nova.compute.manager [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.270038] env[61986]: DEBUG nova.network.neutron [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.271820] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/18f87ccd-ca34-44f6-aa7d-d38397cc479b.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1071.272071] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.275030] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2aec92fc-1a44-474a-906f-5bbf8a246b90 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.283045] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160402, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.284377] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1071.284377] env[61986]: value = "task-1160403" [ 1071.284377] env[61986]: _type = "Task" [ 1071.284377] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.293614] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160403, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.294479] env[61986]: DEBUG nova.network.neutron [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1071.568032] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160401, 'name': ReconfigVM_Task, 'duration_secs': 0.316753} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.568032] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfigured VM instance instance-00000052 to attach disk [datastore2] volume-3f7d5322-8141-4aa7-813d-2ae4f033983e/volume-3f7d5322-8141-4aa7-813d-2ae4f033983e.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.572272] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50bad170-9f06-494c-aaa8-6224c94fc133 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.582133] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.582263] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.582404] env[61986]: DEBUG nova.network.neutron [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1071.589899] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1071.589899] env[61986]: value = "task-1160404" [ 1071.589899] env[61986]: _type = "Task" [ 1071.589899] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.605263] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160404, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.782100] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160402, 'name': ReconfigVM_Task, 'duration_secs': 0.277665} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.782388] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Reconfigured VM instance instance-00000062 to attach disk [datastore2] c3bcd1ba-a3df-4b81-9743-431235536642/c3bcd1ba-a3df-4b81-9743-431235536642.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.783060] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9bc28b7-10cd-4af5-8ac9-ec07a3d95406 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.790818] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1071.790818] env[61986]: value = "task-1160405" [ 1071.790818] env[61986]: _type = "Task" [ 1071.790818] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.793623] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160403, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070682} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.796539] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.796841] env[61986]: DEBUG nova.network.neutron [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.798386] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e52d46-ef60-4053-9d12-f80eacf3598e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.806718] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160405, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.825220] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/18f87ccd-ca34-44f6-aa7d-d38397cc479b.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.825905] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a1b0c9b-8410-4e54-8f6d-0298d6ad0ae1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.846838] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1071.846838] env[61986]: value = "task-1160406" [ 1071.846838] env[61986]: _type = "Task" [ 1071.846838] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.855689] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160406, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.099841] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160404, 'name': ReconfigVM_Task, 'duration_secs': 0.144386} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.100131] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252451', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'name': 'volume-3f7d5322-8141-4aa7-813d-2ae4f033983e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '221e10a3-da31-410c-80f8-4bcc2c515710', 'attached_at': '', 'detached_at': '', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'serial': '3f7d5322-8141-4aa7-813d-2ae4f033983e'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1072.100766] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb4504d6-8184-4174-9e9a-197a9ee91429 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.109057] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1072.109057] env[61986]: value = "task-1160407" [ 1072.109057] env[61986]: _type = "Task" [ 1072.109057] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.115243] env[61986]: DEBUG nova.network.neutron [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1072.120639] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160407, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.249389] env[61986]: DEBUG nova.network.neutron [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updating instance_info_cache with network_info: [{"id": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "address": "fa:16:3e:e3:03:02", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73c2047a-04", "ovs_interfaceid": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.302559] env[61986]: INFO nova.compute.manager [-] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Took 1.03 seconds to deallocate network for instance. [ 1072.302890] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160405, 'name': Rename_Task, 'duration_secs': 0.138435} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.304550] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1072.304792] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-115c8786-a3a7-4de5-925f-417c670064f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.314688] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1072.314688] env[61986]: value = "task-1160408" [ 1072.314688] env[61986]: _type = "Task" [ 1072.314688] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.323212] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160408, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.356548] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160406, 'name': ReconfigVM_Task, 'duration_secs': 0.27173} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.356851] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/18f87ccd-ca34-44f6-aa7d-d38397cc479b.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.357479] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c4172b1-12b5-4124-aca3-3d4810605a10 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.364586] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1072.364586] env[61986]: value = "task-1160409" [ 1072.364586] env[61986]: _type = "Task" [ 1072.364586] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.373053] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160409, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.619220] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160407, 'name': Rename_Task, 'duration_secs': 0.159266} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.619506] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1072.619828] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b6fa42f-4c02-42f0-9d44-d255e2521ff8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.627829] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1072.627829] env[61986]: value = "task-1160410" [ 1072.627829] env[61986]: _type = "Task" [ 1072.627829] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.636633] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.752594] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.752950] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Instance network_info: |[{"id": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "address": "fa:16:3e:e3:03:02", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73c2047a-04", "ovs_interfaceid": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1072.753754] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:03:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73c2047a-043c-4e3a-ae94-b9749f2229f3', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.763224] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating folder: Project (3d92ccda87d241068595992a1b8d3029). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1072.763563] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e39718f-8f90-44bd-8eec-f71524597ac9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.778173] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created folder: Project (3d92ccda87d241068595992a1b8d3029) in parent group-v252271. [ 1072.778443] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating folder: Instances. Parent ref: group-v252465. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1072.778808] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ca4be13-eba3-4444-97c9-bc383cde90c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.789353] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created folder: Instances in parent group-v252465. [ 1072.789583] env[61986]: DEBUG oslo.service.loopingcall [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.789777] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1072.789988] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfc84f2c-4ed5-4fdc-b8f2-0f44890fdc3f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.808570] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.808570] env[61986]: value = "task-1160413" [ 1072.808570] env[61986]: _type = "Task" [ 1072.808570] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.812294] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.812534] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.812751] env[61986]: DEBUG nova.objects.instance [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lazy-loading 'resources' on Instance uuid fc3ae819-efa0-497e-8d2f-d17a53726056 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.821126] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160413, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.826797] env[61986]: DEBUG oslo_vmware.api [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160408, 'name': PowerOnVM_Task, 'duration_secs': 0.493969} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.827052] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1072.827260] env[61986]: INFO nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Took 10.73 seconds to spawn the instance on the hypervisor. [ 1072.827440] env[61986]: DEBUG nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.828244] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a073cc75-58ad-439a-87a2-8b7baa0092f9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.877017] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160409, 'name': Rename_Task, 'duration_secs': 0.15454} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.877321] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1072.877563] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e3b2ddf-5210-4009-aa8c-481413d49043 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.883951] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1072.883951] env[61986]: value = "task-1160414" [ 1072.883951] env[61986]: _type = "Task" [ 1072.883951] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.894117] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160414, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.902906] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1072.903228] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252464', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'name': 'volume-68570cec-bfcf-4107-82c7-cbbf826dca77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9efa4214-3954-49ef-bc7b-6daba8ba07f2', 'attached_at': '', 'detached_at': '', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'serial': '68570cec-bfcf-4107-82c7-cbbf826dca77'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1072.904074] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70671096-23fe-4049-b4b1-0eb131f5ae9c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.919972] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99b8f60-dc58-401c-98a6-e00b55a6e7a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.944889] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] volume-68570cec-bfcf-4107-82c7-cbbf826dca77/volume-68570cec-bfcf-4107-82c7-cbbf826dca77.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.945511] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bd68b33-ed5d-46c3-9bbb-edf662880c4d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.964178] env[61986]: DEBUG oslo_vmware.api [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1072.964178] env[61986]: value = "task-1160415" [ 1072.964178] env[61986]: _type = "Task" [ 1072.964178] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.972506] env[61986]: DEBUG oslo_vmware.api [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.112870] env[61986]: DEBUG nova.compute.manager [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Received event network-changed-73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1073.113019] env[61986]: DEBUG nova.compute.manager [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Refreshing instance network info cache due to event network-changed-73c2047a-043c-4e3a-ae94-b9749f2229f3. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1073.113228] env[61986]: DEBUG oslo_concurrency.lockutils [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] Acquiring lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.113380] env[61986]: DEBUG oslo_concurrency.lockutils [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] Acquired lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.113541] env[61986]: DEBUG nova.network.neutron [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Refreshing network info cache for port 73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1073.138326] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160410, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.320602] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160413, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.345460] env[61986]: INFO nova.compute.manager [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Took 28.82 seconds to build instance. [ 1073.396634] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160414, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.428957] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912ba57f-a6fe-4cbb-8172-9707ee358592 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.436721] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f624163-ad4d-429d-8c01-0608d701c75d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.471564] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c7313d-b69b-4f69-b686-506e2605ab7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.482516] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade0f30e-2b45-45e3-867c-8ae666781737 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.487262] env[61986]: DEBUG oslo_vmware.api [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160415, 'name': ReconfigVM_Task, 'duration_secs': 0.35116} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.487689] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Reconfigured VM instance instance-00000061 to attach disk [datastore2] volume-68570cec-bfcf-4107-82c7-cbbf826dca77/volume-68570cec-bfcf-4107-82c7-cbbf826dca77.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.494233] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af69fb68-adb2-4c1c-8929-235609b5f4b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.513253] env[61986]: DEBUG nova.compute.provider_tree [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.521591] env[61986]: DEBUG oslo_vmware.api [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1073.521591] env[61986]: value = "task-1160416" [ 1073.521591] env[61986]: _type = "Task" [ 1073.521591] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.533422] env[61986]: DEBUG oslo_vmware.api [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160416, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.640075] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160410, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.820215] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160413, 'name': CreateVM_Task, 'duration_secs': 0.99135} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.820411] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1073.821207] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.821424] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.821826] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.822154] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfe50eba-d71f-4474-8588-7b5fd23047ec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.828166] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1073.828166] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5291efd6-2b46-e0c7-ea7b-773f13e1e9ce" [ 1073.828166] env[61986]: _type = "Task" [ 1073.828166] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.837565] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5291efd6-2b46-e0c7-ea7b-773f13e1e9ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.848007] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ee69e146-2504-400b-a5f4-69a2a04e49dc tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "c3bcd1ba-a3df-4b81-9743-431235536642" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.329s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.894919] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160414, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.006806] env[61986]: DEBUG nova.network.neutron [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updated VIF entry in instance network info cache for port 73c2047a-043c-4e3a-ae94-b9749f2229f3. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1074.007357] env[61986]: DEBUG nova.network.neutron [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updating instance_info_cache with network_info: [{"id": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "address": "fa:16:3e:e3:03:02", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73c2047a-04", "ovs_interfaceid": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.015845] env[61986]: DEBUG nova.scheduler.client.report [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.034373] env[61986]: DEBUG oslo_vmware.api [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160416, 'name': ReconfigVM_Task, 'duration_secs': 0.143978} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.035589] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252464', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'name': 'volume-68570cec-bfcf-4107-82c7-cbbf826dca77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9efa4214-3954-49ef-bc7b-6daba8ba07f2', 'attached_at': '', 'detached_at': '', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'serial': '68570cec-bfcf-4107-82c7-cbbf826dca77'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1074.142413] env[61986]: DEBUG oslo_vmware.api [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160410, 'name': PowerOnVM_Task, 'duration_secs': 1.359207} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.142577] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1074.268072] env[61986]: DEBUG nova.compute.manager [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.269312] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7095d6-9e39-4b57-9a06-0dbf0a1a970b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.339224] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5291efd6-2b46-e0c7-ea7b-773f13e1e9ce, 'name': SearchDatastore_Task, 'duration_secs': 0.011346} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.339606] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.339756] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1074.339999] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.340174] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.340362] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1074.340624] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f959c47-ee65-4979-9c91-3047aa519e55 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.353265] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1074.354099] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1074.354212] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f3c8859-c439-48c8-a365-575b340c8e64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.359945] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1074.359945] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520b3656-db36-9607-4e62-ad3557bfd999" [ 1074.359945] env[61986]: _type = "Task" [ 1074.359945] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.367681] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520b3656-db36-9607-4e62-ad3557bfd999, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.394819] env[61986]: DEBUG oslo_vmware.api [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160414, 'name': PowerOnVM_Task, 'duration_secs': 1.080128} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.395034] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1074.395253] env[61986]: INFO nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Took 10.04 seconds to spawn the instance on the hypervisor. [ 1074.395434] env[61986]: DEBUG nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.396186] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a096b7c-b7b7-4922-968c-6a3f2eff8bda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.509893] env[61986]: DEBUG oslo_concurrency.lockutils [req-6222bcaf-44e4-4fce-83f1-fb438dace086 req-992df259-4195-4c52-bdfe-2349dca5c8a9 service nova] Releasing lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.523215] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.548036] env[61986]: INFO nova.scheduler.client.report [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Deleted allocations for instance fc3ae819-efa0-497e-8d2f-d17a53726056 [ 1074.597105] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "c3bcd1ba-a3df-4b81-9743-431235536642" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.597276] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "c3bcd1ba-a3df-4b81-9743-431235536642" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.597493] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "c3bcd1ba-a3df-4b81-9743-431235536642-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.597683] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "c3bcd1ba-a3df-4b81-9743-431235536642-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.597941] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "c3bcd1ba-a3df-4b81-9743-431235536642-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.600453] env[61986]: INFO nova.compute.manager [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Terminating instance [ 1074.602261] env[61986]: DEBUG nova.compute.manager [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1074.602461] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1074.603296] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a4ac17-9e6a-4678-9409-66c3a9dee67a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.611679] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1074.612055] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-710628a6-d733-42e0-bb04-0c22a722ba75 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.618318] env[61986]: DEBUG oslo_vmware.api [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1074.618318] env[61986]: value = "task-1160417" [ 1074.618318] env[61986]: _type = "Task" [ 1074.618318] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.626582] env[61986]: DEBUG oslo_vmware.api [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.787265] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f16ec49a-8e43-44e3-8a50-d59cde05e8f7 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 40.775s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.872324] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520b3656-db36-9607-4e62-ad3557bfd999, 'name': SearchDatastore_Task, 'duration_secs': 0.032191} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.873196] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e79dde4-e414-49f4-a58f-f89b0e1b9223 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.881688] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1074.881688] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c777b-8e41-da77-e631-cd3a10d9a111" [ 1074.881688] env[61986]: _type = "Task" [ 1074.881688] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.889812] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c777b-8e41-da77-e631-cd3a10d9a111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.914911] env[61986]: INFO nova.compute.manager [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Took 22.77 seconds to build instance. [ 1075.057479] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6477409a-0ae2-4480-8796-c757c25f625e tempest-ServerShowV257Test-720064922 tempest-ServerShowV257Test-720064922-project-member] Lock "fc3ae819-efa0-497e-8d2f-d17a53726056" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.490s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.077445] env[61986]: DEBUG nova.objects.instance [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'flavor' on Instance uuid 9efa4214-3954-49ef-bc7b-6daba8ba07f2 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.131776] env[61986]: DEBUG oslo_vmware.api [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160417, 'name': PowerOffVM_Task, 'duration_secs': 0.255455} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.132082] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1075.132261] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1075.132587] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1617cb4a-c55c-4f55-9fb3-1cde8f717145 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.229916] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1075.230197] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1075.230470] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleting the datastore file [datastore2] c3bcd1ba-a3df-4b81-9743-431235536642 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.230903] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ac79063-8754-4c82-b8b2-0668a97bd00b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.238572] env[61986]: DEBUG oslo_vmware.api [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for the task: (returnval){ [ 1075.238572] env[61986]: value = "task-1160419" [ 1075.238572] env[61986]: _type = "Task" [ 1075.238572] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.247361] env[61986]: DEBUG oslo_vmware.api [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.396120] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525c777b-8e41-da77-e631-cd3a10d9a111, 'name': SearchDatastore_Task, 'duration_secs': 0.01113} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.396530] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.396873] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350/487d0b5e-27e6-4e9c-87b1-81e0aa9dd350.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1075.397262] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b5e360d-c81a-4e46-b46c-d095ef479f26 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.407766] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1075.407766] env[61986]: value = "task-1160420" [ 1075.407766] env[61986]: _type = "Task" [ 1075.407766] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.417832] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fdd45c48-796c-4578-ad81-32e6b88bbdaf tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.285s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.418133] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160420, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.583215] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b566e1d9-9c7d-440a-999d-78a4608d6a89 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.296s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.660351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.660618] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.660878] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.661087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.661267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.668751] env[61986]: INFO nova.compute.manager [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Terminating instance [ 1075.671857] env[61986]: DEBUG nova.compute.manager [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1075.672094] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1075.672381] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f03875d-08da-40e1-ba2a-aa77c11dc4d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.685008] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1075.685008] env[61986]: value = "task-1160421" [ 1075.685008] env[61986]: _type = "Task" [ 1075.685008] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.695130] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160421, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.750203] env[61986]: DEBUG oslo_vmware.api [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Task: {'id': task-1160419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329173} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.750490] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.750755] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1075.750970] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1075.751172] env[61986]: INFO nova.compute.manager [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1075.751435] env[61986]: DEBUG oslo.service.loopingcall [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.751690] env[61986]: DEBUG nova.compute.manager [-] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1075.751837] env[61986]: DEBUG nova.network.neutron [-] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1075.920972] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160420, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.029647] env[61986]: INFO nova.compute.manager [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Rescuing [ 1076.029982] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.030218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.030457] env[61986]: DEBUG nova.network.neutron [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1076.142990] env[61986]: DEBUG nova.compute.manager [req-51770ae8-6ac7-44ab-afe1-e601c4e25666 req-43158afe-d3f6-43b9-b1a6-8b16be7927ae service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Received event network-vif-deleted-ae574939-8249-4b31-b7b3-239d0d191f4b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1076.143227] env[61986]: INFO nova.compute.manager [req-51770ae8-6ac7-44ab-afe1-e601c4e25666 req-43158afe-d3f6-43b9-b1a6-8b16be7927ae service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Neutron deleted interface ae574939-8249-4b31-b7b3-239d0d191f4b; detaching it from the instance and deleting it from the info cache [ 1076.143407] env[61986]: DEBUG nova.network.neutron [req-51770ae8-6ac7-44ab-afe1-e601c4e25666 req-43158afe-d3f6-43b9-b1a6-8b16be7927ae service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.194875] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160421, 'name': PowerOffVM_Task, 'duration_secs': 0.246808} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.195165] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1076.195362] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1076.195556] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252464', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'name': 'volume-68570cec-bfcf-4107-82c7-cbbf826dca77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9efa4214-3954-49ef-bc7b-6daba8ba07f2', 'attached_at': '', 'detached_at': '', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'serial': '68570cec-bfcf-4107-82c7-cbbf826dca77'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1076.196350] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec5e82a-df7d-4bb1-8f4d-21fdb249dec7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.216976] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57fee8e-9887-4cfc-bdf4-f63fcd65183a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.223660] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394e36f0-0491-4a88-b9e4-d12ecea4a6e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.243381] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12946926-d234-43a3-9411-84270dee24f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.257457] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] The volume has not been displaced from its original location: [datastore2] volume-68570cec-bfcf-4107-82c7-cbbf826dca77/volume-68570cec-bfcf-4107-82c7-cbbf826dca77.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1076.263083] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1076.263352] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fdb6fbb-1721-49e8-b227-1591e663e941 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.281535] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1076.281535] env[61986]: value = "task-1160422" [ 1076.281535] env[61986]: _type = "Task" [ 1076.281535] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.289501] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.420426] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160420, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656807} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.420815] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350/487d0b5e-27e6-4e9c-87b1-81e0aa9dd350.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1076.421120] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1076.421385] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd14f8b1-78ca-4121-ae6a-e12cb131f214 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.429220] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1076.429220] env[61986]: value = "task-1160423" [ 1076.429220] env[61986]: _type = "Task" [ 1076.429220] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.440221] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.616472] env[61986]: DEBUG nova.network.neutron [-] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.646600] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-568321db-25bd-46a1-ac30-68e75c9ea8a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.657694] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31816a41-12a4-4a57-93b4-63a6d67e4bf9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.689839] env[61986]: DEBUG nova.compute.manager [req-51770ae8-6ac7-44ab-afe1-e601c4e25666 req-43158afe-d3f6-43b9-b1a6-8b16be7927ae service nova] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Detach interface failed, port_id=ae574939-8249-4b31-b7b3-239d0d191f4b, reason: Instance c3bcd1ba-a3df-4b81-9743-431235536642 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1076.761521] env[61986]: DEBUG nova.network.neutron [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.791920] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160422, 'name': ReconfigVM_Task, 'duration_secs': 0.203588} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.792224] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1076.796854] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90543dd1-2ca1-4169-9b38-0733c281891d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.813061] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1076.813061] env[61986]: value = "task-1160424" [ 1076.813061] env[61986]: _type = "Task" [ 1076.813061] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.821171] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160424, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.939328] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07705} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.939619] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1076.940420] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274fd576-9419-462e-b50d-339f70f795c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.963141] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350/487d0b5e-27e6-4e9c-87b1-81e0aa9dd350.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.963449] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25b36c40-b957-4131-aa54-75c5a7b9c10d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.984912] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1076.984912] env[61986]: value = "task-1160425" [ 1076.984912] env[61986]: _type = "Task" [ 1076.984912] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.993348] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160425, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.119861] env[61986]: INFO nova.compute.manager [-] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Took 1.37 seconds to deallocate network for instance. [ 1077.264016] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.325452] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160424, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.497345] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160425, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.598450] env[61986]: DEBUG oslo_concurrency.lockutils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.599883] env[61986]: DEBUG oslo_concurrency.lockutils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.626365] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.626644] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.626869] env[61986]: DEBUG nova.objects.instance [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lazy-loading 'resources' on Instance uuid c3bcd1ba-a3df-4b81-9743-431235536642 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.796767] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1077.797040] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3035607c-42d7-4e96-8755-2af5afff602a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.804726] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1077.804726] env[61986]: value = "task-1160426" [ 1077.804726] env[61986]: _type = "Task" [ 1077.804726] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.815087] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160426, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.822865] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160424, 'name': ReconfigVM_Task, 'duration_secs': 0.616517} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.823164] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252464', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'name': 'volume-68570cec-bfcf-4107-82c7-cbbf826dca77', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9efa4214-3954-49ef-bc7b-6daba8ba07f2', 'attached_at': '', 'detached_at': '', 'volume_id': '68570cec-bfcf-4107-82c7-cbbf826dca77', 'serial': '68570cec-bfcf-4107-82c7-cbbf826dca77'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1077.823461] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1077.824210] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddce70a8-92e3-4c66-80a9-253c17d33150 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.830595] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1077.830841] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-494ca0f8-ea7d-49bd-850f-86a611863c8c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.905679] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1077.905895] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1077.906101] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleting the datastore file [datastore2] 9efa4214-3954-49ef-bc7b-6daba8ba07f2 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.906439] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66348106-57d6-41f0-b1ab-e4beb8af6a5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.913825] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1077.913825] env[61986]: value = "task-1160428" [ 1077.913825] env[61986]: _type = "Task" [ 1077.913825] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.922421] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.996098] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160425, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.102537] env[61986]: DEBUG nova.compute.utils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.218732] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b60a7c-904a-4c66-b933-1e76ed60f065 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.226172] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe858463-6d51-4ede-9bc0-bc18303f43b0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.255551] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69551aa9-ffed-4ae2-90ea-bbf23820e685 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.262557] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea24bf56-3622-4742-a2bb-2e67e4e0434b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.276653] env[61986]: DEBUG nova.compute.provider_tree [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.313976] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160426, 'name': PowerOffVM_Task, 'duration_secs': 0.241655} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.314265] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1078.315041] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20fd913-dcce-4a4b-bca6-93eb83d64e36 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.333231] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff06b707-10eb-4424-ac83-df65b850715b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.359740] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.360062] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a4d3ca2-2ce9-47a3-87b5-8cd86aa66887 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.367719] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1078.367719] env[61986]: value = "task-1160429" [ 1078.367719] env[61986]: _type = "Task" [ 1078.367719] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.375533] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.423977] env[61986]: DEBUG oslo_vmware.api [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147012} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.424277] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.424446] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1078.424625] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1078.424801] env[61986]: INFO nova.compute.manager [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Took 2.75 seconds to destroy the instance on the hypervisor. [ 1078.425264] env[61986]: DEBUG oslo.service.loopingcall [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.425510] env[61986]: DEBUG nova.compute.manager [-] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1078.425608] env[61986]: DEBUG nova.network.neutron [-] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1078.495274] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160425, 'name': ReconfigVM_Task, 'duration_secs': 1.038326} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.495675] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350/487d0b5e-27e6-4e9c-87b1-81e0aa9dd350.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.496219] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eed9b9cb-1995-443b-a616-a9b033e87846 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.503191] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1078.503191] env[61986]: value = "task-1160430" [ 1078.503191] env[61986]: _type = "Task" [ 1078.503191] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.511439] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160430, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.605154] env[61986]: DEBUG oslo_concurrency.lockutils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.710326] env[61986]: DEBUG nova.compute.manager [req-8122c0ec-1168-459a-b1c3-a2155cce8144 req-5f71fe35-ef84-40e5-92ee-ef62e70e3254 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Received event network-vif-deleted-b638bc20-c1b8-4047-a795-2d463953e54a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1078.710533] env[61986]: INFO nova.compute.manager [req-8122c0ec-1168-459a-b1c3-a2155cce8144 req-5f71fe35-ef84-40e5-92ee-ef62e70e3254 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Neutron deleted interface b638bc20-c1b8-4047-a795-2d463953e54a; detaching it from the instance and deleting it from the info cache [ 1078.710748] env[61986]: DEBUG nova.network.neutron [req-8122c0ec-1168-459a-b1c3-a2155cce8144 req-5f71fe35-ef84-40e5-92ee-ef62e70e3254 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.779087] env[61986]: DEBUG nova.scheduler.client.report [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.880035] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1078.880350] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.880712] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.880911] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.881182] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.881505] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-690ef6ab-aa11-4058-850e-52f329fa2860 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.892956] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.893191] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1078.894186] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be85ca0e-2311-4251-864a-d48966591748 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.900550] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1078.900550] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5282f20a-c7b4-a66a-ead8-5dabef35faf5" [ 1078.900550] env[61986]: _type = "Task" [ 1078.900550] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.910384] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5282f20a-c7b4-a66a-ead8-5dabef35faf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.013524] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160430, 'name': Rename_Task, 'duration_secs': 0.143116} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.013813] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1079.014078] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54b4f9b8-4bc6-48de-8edd-dce39f41762e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.020796] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1079.020796] env[61986]: value = "task-1160431" [ 1079.020796] env[61986]: _type = "Task" [ 1079.020796] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.028298] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160431, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.155707] env[61986]: DEBUG nova.network.neutron [-] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.212986] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1059209-7c4d-4814-a618-920e70d46f59 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.223448] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36096cf-c661-40e1-830d-b884ca9bccca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.253421] env[61986]: DEBUG nova.compute.manager [req-8122c0ec-1168-459a-b1c3-a2155cce8144 req-5f71fe35-ef84-40e5-92ee-ef62e70e3254 service nova] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Detach interface failed, port_id=b638bc20-c1b8-4047-a795-2d463953e54a, reason: Instance 9efa4214-3954-49ef-bc7b-6daba8ba07f2 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1079.284693] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.305594] env[61986]: INFO nova.scheduler.client.report [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Deleted allocations for instance c3bcd1ba-a3df-4b81-9743-431235536642 [ 1079.410884] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5282f20a-c7b4-a66a-ead8-5dabef35faf5, 'name': SearchDatastore_Task, 'duration_secs': 0.00824} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.411790] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12914100-b1c9-4300-8ce2-46d2b8fb36d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.419234] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1079.419234] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b03570-d139-f6b5-4335-f7c16eb32ba1" [ 1079.419234] env[61986]: _type = "Task" [ 1079.419234] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.428700] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b03570-d139-f6b5-4335-f7c16eb32ba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.531187] env[61986]: DEBUG oslo_vmware.api [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160431, 'name': PowerOnVM_Task, 'duration_secs': 0.429502} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.531461] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1079.531657] env[61986]: INFO nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Took 8.96 seconds to spawn the instance on the hypervisor. [ 1079.531861] env[61986]: DEBUG nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1079.532652] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74aad4c-944e-40ee-a28c-6a0b0d8e96c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.658013] env[61986]: INFO nova.compute.manager [-] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Took 1.23 seconds to deallocate network for instance. [ 1079.671281] env[61986]: DEBUG oslo_concurrency.lockutils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.671465] env[61986]: DEBUG oslo_concurrency.lockutils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.671701] env[61986]: INFO nova.compute.manager [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Attaching volume 5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5 to /dev/sdb [ 1079.705219] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1c1b2f-9da0-4977-be49-84a1ff495635 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.713374] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c4eaeb-2dd6-42d4-bd54-fd5832e57dfe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.727727] env[61986]: DEBUG nova.virt.block_device [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updating existing volume attachment record: ef254c03-136d-4e42-b91b-31f066a5bf8d {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1079.813432] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cf76a923-5c0d-4f8a-abd6-4ce5a32aec50 tempest-ServerDiskConfigTestJSON-1563111558 tempest-ServerDiskConfigTestJSON-1563111558-project-member] Lock "c3bcd1ba-a3df-4b81-9743-431235536642" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.216s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.930101] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b03570-d139-f6b5-4335-f7c16eb32ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.010658} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.930413] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.930864] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk. {{(pid=61986) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1079.931039] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e733d38-90ae-4ee1-ab4a-d595957ff1ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.939337] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1079.939337] env[61986]: value = "task-1160433" [ 1079.939337] env[61986]: _type = "Task" [ 1079.939337] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.948712] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160433, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.052658] env[61986]: INFO nova.compute.manager [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Took 13.73 seconds to build instance. [ 1080.205938] env[61986]: INFO nova.compute.manager [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Took 0.55 seconds to detach 1 volumes for instance. [ 1080.450804] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160433, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469207} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.451076] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk. [ 1080.451855] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b03378-e119-4c99-9fbf-1122c85f55cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.476866] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.477154] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e46b1ef0-5bbd-450a-a322-c86a5d3ec20f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.495855] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1080.495855] env[61986]: value = "task-1160434" [ 1080.495855] env[61986]: _type = "Task" [ 1080.495855] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.504083] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160434, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.554509] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cefeb453-2e9d-4957-8aaa-0976974f906c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.238s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.716571] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.716571] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.716571] env[61986]: DEBUG nova.objects.instance [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'resources' on Instance uuid 9efa4214-3954-49ef-bc7b-6daba8ba07f2 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.008339] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160434, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.321779] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad7fd98-750f-4261-bdee-0a704c19b44e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.336334] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0acc416d-15f8-4ecb-ac17-ec2efdd1aaed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.377341] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9888fff6-be50-47f1-afc4-0b64b51ec519 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.385933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2148aa9-8648-49f3-97ce-54de21f05ed7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.402998] env[61986]: DEBUG nova.compute.provider_tree [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.509577] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160434, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.906491] env[61986]: DEBUG nova.scheduler.client.report [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.012275] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160434, 'name': ReconfigVM_Task, 'duration_secs': 1.192736} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.012593] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5-rescue.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.013444] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9955e88-f178-4df3-ab1e-2676cb7fafa4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.039038] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fefe71ae-9ad6-4bbc-8954-e7142eb27ad0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.055734] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1082.055734] env[61986]: value = "task-1160436" [ 1082.055734] env[61986]: _type = "Task" [ 1082.055734] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.064857] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160436, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.372519] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.372773] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.411292] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.437694] env[61986]: INFO nova.scheduler.client.report [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocations for instance 9efa4214-3954-49ef-bc7b-6daba8ba07f2 [ 1082.566662] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160436, 'name': ReconfigVM_Task, 'duration_secs': 0.160135} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.566831] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1082.567238] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-79cee495-aa18-4992-9ad7-fc6d4e4d3c5f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.574235] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1082.574235] env[61986]: value = "task-1160437" [ 1082.574235] env[61986]: _type = "Task" [ 1082.574235] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.582864] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.653204] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.653449] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.875821] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1082.946635] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f9f17821-f110-447c-811f-384bec27d517 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9efa4214-3954-49ef-bc7b-6daba8ba07f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.286s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.086697] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160437, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.156359] env[61986]: DEBUG nova.compute.utils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1083.401912] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.406229] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.004s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.407849] env[61986]: INFO nova.compute.claims [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1083.589029] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160437, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.659122] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.086887] env[61986]: DEBUG oslo_vmware.api [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160437, 'name': PowerOnVM_Task, 'duration_secs': 1.040724} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.087381] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1084.090096] env[61986]: DEBUG nova.compute.manager [None req-c6e9bd0d-943d-4ae4-856f-9ba9e79e1062 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1084.090892] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fb010d-0390-418e-aec1-5c4ac1ac8c69 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.272484] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1084.272844] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252468', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'name': 'volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31ba6f-0a3e-4a1c-af5e-617e8843839f', 'attached_at': '', 'detached_at': '', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'serial': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1084.273645] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94764ba5-8a64-4392-af97-9b42cb3fb76e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.294589] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3767639e-b944-464e-bc9b-3cd22a62a87c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.324491] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5/volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.324640] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54929224-9161-400b-8482-d5e4d9cf73a6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.344197] env[61986]: DEBUG oslo_vmware.api [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1084.344197] env[61986]: value = "task-1160438" [ 1084.344197] env[61986]: _type = "Task" [ 1084.344197] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.352957] env[61986]: DEBUG oslo_vmware.api [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.537786] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b864f106-4ef8-45ea-ab7b-b380a2fa89ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.545249] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.545347] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.551351] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18adbd24-69e2-4785-957e-d1d84d719cee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.596402] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7810eb44-842e-4dfd-9052-95942cd194cb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.606836] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05985b13-72cf-4431-bf6d-fca0364efe96 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.624949] env[61986]: DEBUG nova.compute.provider_tree [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.729180] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.729472] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.729718] env[61986]: INFO nova.compute.manager [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Attaching volume 1568453a-0d54-4730-b64d-cb77ca01e291 to /dev/sdb [ 1084.770317] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c9fa5d-6f53-4c32-8823-6dc7e52f5892 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.777854] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d029ec-3a04-4e3f-bd06-f49423a73a1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.792779] env[61986]: DEBUG nova.virt.block_device [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating existing volume attachment record: 5b71aef4-03c4-45ac-8615-85287f958035 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1084.859817] env[61986]: DEBUG oslo_vmware.api [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160438, 'name': ReconfigVM_Task, 'duration_secs': 0.450541} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.860220] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5/volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.865056] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b94346f5-b162-4438-bccc-dad49d7050d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.882509] env[61986]: DEBUG oslo_vmware.api [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1084.882509] env[61986]: value = "task-1160439" [ 1084.882509] env[61986]: _type = "Task" [ 1084.882509] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.891099] env[61986]: DEBUG oslo_vmware.api [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160439, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.048480] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1085.132087] env[61986]: DEBUG nova.scheduler.client.report [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.393350] env[61986]: DEBUG oslo_vmware.api [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160439, 'name': ReconfigVM_Task, 'duration_secs': 0.169558} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.393691] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252468', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'name': 'volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31ba6f-0a3e-4a1c-af5e-617e8843839f', 'attached_at': '', 'detached_at': '', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'serial': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1085.574717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.639447] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.640146] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1085.643366] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.069s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.644834] env[61986]: INFO nova.compute.claims [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.862094] env[61986]: DEBUG nova.compute.manager [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.862306] env[61986]: DEBUG nova.compute.manager [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing instance network info cache due to event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1085.862522] env[61986]: DEBUG oslo_concurrency.lockutils [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.862701] env[61986]: DEBUG oslo_concurrency.lockutils [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.862826] env[61986]: DEBUG nova.network.neutron [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.149355] env[61986]: DEBUG nova.compute.utils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1086.153097] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1086.153275] env[61986]: DEBUG nova.network.neutron [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1086.213888] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.214072] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1086.229730] env[61986]: DEBUG nova.policy [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1086.309687] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "a6094cc4-7d17-4858-abae-972425241c8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.309687] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "a6094cc4-7d17-4858-abae-972425241c8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.434127] env[61986]: DEBUG nova.objects.instance [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'flavor' on Instance uuid af31ba6f-0a3e-4a1c-af5e-617e8843839f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.572988] env[61986]: DEBUG nova.network.neutron [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updated VIF entry in instance network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1086.572988] env[61986]: DEBUG nova.network.neutron [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.658462] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1086.690758] env[61986]: DEBUG nova.network.neutron [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Successfully created port: 6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1086.804229] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bff270b-9d83-4e08-8162-d626941abe3f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.813098] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1086.816678] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0843a85-1f74-4976-b7c6-66db472a808f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.853222] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcd59f8-5821-4f0a-91eb-6cd845ab8c2b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.865474] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5458d4-a07a-4887-a1fa-d83e711b8b80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.880789] env[61986]: DEBUG nova.compute.provider_tree [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.940190] env[61986]: DEBUG oslo_concurrency.lockutils [None req-61f5ff61-526a-41d4-9367-4971240ab67b tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.269s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.076374] env[61986]: DEBUG oslo_concurrency.lockutils [req-7b9c26a5-c2d1-41d0-85b6-41af2120dd3f req-2d54711d-c0ca-490d-9516-c298b074051c service nova] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.336643] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.362293] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.362965] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.003s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.384206] env[61986]: DEBUG nova.scheduler.client.report [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.669065] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1087.694418] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1087.694677] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1087.694838] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.695037] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1087.695192] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.695342] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1087.695552] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1087.695719] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1087.695923] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1087.696121] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1087.696302] env[61986]: DEBUG nova.virt.hardware [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1087.697200] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a1ec5d-4c2e-43ea-b122-3b52b542228a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.705373] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6528594-6b87-4f0c-b62b-4a12a67e87d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.866587] env[61986]: INFO nova.compute.manager [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Detaching volume 5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5 [ 1087.893351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.893351] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1087.898017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.559s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.898017] env[61986]: INFO nova.compute.claims [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.900270] env[61986]: DEBUG nova.compute.manager [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.900509] env[61986]: DEBUG nova.compute.manager [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing instance network info cache due to event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1087.900796] env[61986]: DEBUG oslo_concurrency.lockutils [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.901102] env[61986]: DEBUG oslo_concurrency.lockutils [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.901299] env[61986]: DEBUG nova.network.neutron [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1087.908530] env[61986]: INFO nova.virt.block_device [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Attempting to driver detach volume 5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5 from mountpoint /dev/sdb [ 1087.908805] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.909067] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252468', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'name': 'volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31ba6f-0a3e-4a1c-af5e-617e8843839f', 'attached_at': '', 'detached_at': '', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'serial': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.909941] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443496b4-17f9-400c-97b5-12850bbeefd1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.936134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83dad77-6884-42b9-ae5a-877253953a21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.945305] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03bc8fa-f854-431b-8ff5-5cf2135b4865 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.968125] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda43e28-7fd9-436f-aaaf-87ce7ee8d38b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.987093] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] The volume has not been displaced from its original location: [datastore2] volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5/volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1087.992969] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1087.993315] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec2cc62b-6fe1-4c23-a838-be1df4364928 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.014338] env[61986]: DEBUG oslo_vmware.api [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1088.014338] env[61986]: value = "task-1160444" [ 1088.014338] env[61986]: _type = "Task" [ 1088.014338] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.023197] env[61986]: DEBUG oslo_vmware.api [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160444, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.213810] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1088.404496] env[61986]: DEBUG nova.compute.utils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1088.412106] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1088.412297] env[61986]: DEBUG nova.network.neutron [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1088.501284] env[61986]: DEBUG nova.policy [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c40c9363dc7741a3bd6a040f20284837', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e0f7d6b6be498eaec797aa64aaea7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1088.527511] env[61986]: DEBUG oslo_vmware.api [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160444, 'name': ReconfigVM_Task, 'duration_secs': 0.228641} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.527788] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.532583] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d70d5a8-e898-464a-aec4-1216416be926 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.559316] env[61986]: DEBUG oslo_vmware.api [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1088.559316] env[61986]: value = "task-1160445" [ 1088.559316] env[61986]: _type = "Task" [ 1088.559316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.570038] env[61986]: DEBUG oslo_vmware.api [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160445, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.717374] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.754136] env[61986]: DEBUG nova.network.neutron [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updated VIF entry in instance network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1088.755139] env[61986]: DEBUG nova.network.neutron [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.866752] env[61986]: DEBUG nova.network.neutron [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Successfully updated port: 6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1088.868442] env[61986]: DEBUG nova.network.neutron [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Successfully created port: 78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1088.912855] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1089.069573] env[61986]: DEBUG oslo_vmware.api [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160445, 'name': ReconfigVM_Task, 'duration_secs': 0.162909} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.070538] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252468', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'name': 'volume-5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31ba6f-0a3e-4a1c-af5e-617e8843839f', 'attached_at': '', 'detached_at': '', 'volume_id': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5', 'serial': '5a1a5d5e-a72b-4cf4-b5a5-9c7a5acc3aa5'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1089.074801] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57815bc0-4398-4a8f-ae2a-970603ef3b78 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.080952] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d64afc-f295-4431-b8b8-80bcc8f23086 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.115135] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c04147f-b549-41c7-afb2-bb526acdffbc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.123592] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f75680-04c2-4fb7-8fc1-5fa20f60bfbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.136578] env[61986]: DEBUG nova.compute.provider_tree [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.257188] env[61986]: DEBUG oslo_concurrency.lockutils [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.257487] env[61986]: DEBUG nova.compute.manager [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1089.257662] env[61986]: DEBUG nova.compute.manager [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing instance network info cache due to event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1089.257882] env[61986]: DEBUG oslo_concurrency.lockutils [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.258037] env[61986]: DEBUG oslo_concurrency.lockutils [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.258212] env[61986]: DEBUG nova.network.neutron [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1089.373535] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-785dc1af-bec0-447b-90f5-2cb1f53bb0f0" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.373713] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-785dc1af-bec0-447b-90f5-2cb1f53bb0f0" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.373901] env[61986]: DEBUG nova.network.neutron [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1089.617416] env[61986]: DEBUG nova.objects.instance [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'flavor' on Instance uuid af31ba6f-0a3e-4a1c-af5e-617e8843839f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.639751] env[61986]: DEBUG nova.scheduler.client.report [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.838309] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1089.838558] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252470', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'name': 'volume-1568453a-0d54-4730-b64d-cb77ca01e291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e825bf09-7f6c-45db-b6cf-9f2dabb98677', 'attached_at': '', 'detached_at': '', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'serial': '1568453a-0d54-4730-b64d-cb77ca01e291'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1089.839542] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a1578e-4a2c-49e5-9a3e-db78fd181cfb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.857321] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d03e3a-bee9-4692-a550-0ef1e2a4794a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.882133] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-1568453a-0d54-4730-b64d-cb77ca01e291/volume-1568453a-0d54-4730-b64d-cb77ca01e291.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.884386] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db211155-35ed-4159-aeaf-90ee6af60853 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.912276] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1089.912276] env[61986]: value = "task-1160446" [ 1089.912276] env[61986]: _type = "Task" [ 1089.912276] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.922254] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1089.924438] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160446, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.927440] env[61986]: DEBUG nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1089.927664] env[61986]: DEBUG nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing instance network info cache due to event network-changed-12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1089.928059] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Acquiring lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.952301] env[61986]: DEBUG nova.network.neutron [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.957864] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.958262] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.958262] env[61986]: DEBUG nova.virt.hardware [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.959182] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe81034-d553-461c-9d5e-966197aef02f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.966806] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773f6895-4986-4839-a814-c1a688c25d24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.985740] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.985950] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.986179] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.986368] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.986579] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.988732] env[61986]: INFO nova.compute.manager [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Terminating instance [ 1089.990604] env[61986]: DEBUG nova.compute.manager [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.990814] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.991620] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148e9f35-aae8-4a98-9add-19ae8e8fb116 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.000832] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1090.002989] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d4c9498-ae57-4799-ac15-b5823cdef96c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.009490] env[61986]: DEBUG oslo_vmware.api [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1090.009490] env[61986]: value = "task-1160447" [ 1090.009490] env[61986]: _type = "Task" [ 1090.009490] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.018929] env[61986]: DEBUG oslo_vmware.api [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.073831] env[61986]: DEBUG nova.network.neutron [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updated VIF entry in instance network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1090.074236] env[61986]: DEBUG nova.network.neutron [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.131244] env[61986]: DEBUG nova.network.neutron [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Updating instance_info_cache with network_info: [{"id": "6caa2d0a-296d-4ce8-8928-b0b6a4691e39", "address": "fa:16:3e:c3:4d:70", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6caa2d0a-29", "ovs_interfaceid": "6caa2d0a-296d-4ce8-8928-b0b6a4691e39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.144679] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.144679] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1090.147654] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.431s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.147910] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.148150] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1090.149378] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3e77ac-40f9-4ed8-8e77-2a9ac2b69e6e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.158673] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ab2027-6d37-498c-aa4d-943f102c38bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.175125] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdccae87-eeb9-4fce-9bd0-66b9e1307c49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.183886] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a591a79b-56cb-46d5-8347-b4c3bdc03608 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.216394] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181032MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1090.216394] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.216394] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.311713] env[61986]: DEBUG nova.compute.manager [req-f5a76037-8cd1-49cf-a2eb-e2d427269c0c req-6e4a5a6d-9efd-4c82-bf18-2ecc2d60d235 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Received event network-vif-plugged-78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1090.312059] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5a76037-8cd1-49cf-a2eb-e2d427269c0c req-6e4a5a6d-9efd-4c82-bf18-2ecc2d60d235 service nova] Acquiring lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.312459] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5a76037-8cd1-49cf-a2eb-e2d427269c0c req-6e4a5a6d-9efd-4c82-bf18-2ecc2d60d235 service nova] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.312721] env[61986]: DEBUG oslo_concurrency.lockutils [req-f5a76037-8cd1-49cf-a2eb-e2d427269c0c req-6e4a5a6d-9efd-4c82-bf18-2ecc2d60d235 service nova] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.312849] env[61986]: DEBUG nova.compute.manager [req-f5a76037-8cd1-49cf-a2eb-e2d427269c0c req-6e4a5a6d-9efd-4c82-bf18-2ecc2d60d235 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] No waiting events found dispatching network-vif-plugged-78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.313298] env[61986]: WARNING nova.compute.manager [req-f5a76037-8cd1-49cf-a2eb-e2d427269c0c req-6e4a5a6d-9efd-4c82-bf18-2ecc2d60d235 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Received unexpected event network-vif-plugged-78ca1d3d-a907-4b97-8eec-a75b313ca796 for instance with vm_state building and task_state spawning. [ 1090.422698] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160446, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.522107] env[61986]: DEBUG oslo_vmware.api [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160447, 'name': PowerOffVM_Task, 'duration_secs': 0.264759} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.522107] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.522107] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1090.522107] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90f71b35-f95e-467f-923c-354d9d24d2fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.578252] env[61986]: DEBUG oslo_concurrency.lockutils [req-e2adf4a6-bfb7-486f-a8c6-db0db3995d81 req-6142dee5-c538-481a-80a4-f4bd525b9d42 service nova] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.579447] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Acquired lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.579720] env[61986]: DEBUG nova.network.neutron [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Refreshing network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1090.580887] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.581100] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.581293] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Deleting the datastore file [datastore2] 18f87ccd-ca34-44f6-aa7d-d38397cc479b {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.581791] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5db0908-79ec-40a7-8fad-12233d0f0590 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.589103] env[61986]: DEBUG oslo_vmware.api [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for the task: (returnval){ [ 1090.589103] env[61986]: value = "task-1160449" [ 1090.589103] env[61986]: _type = "Task" [ 1090.589103] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.592275] env[61986]: DEBUG nova.network.neutron [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Successfully updated port: 78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.598816] env[61986]: DEBUG oslo_vmware.api [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160449, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.626531] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de4e8f69-ccfc-44d1-bfe2-7f27d3952222 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.264s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.636732] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-785dc1af-bec0-447b-90f5-2cb1f53bb0f0" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.637044] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Instance network_info: |[{"id": "6caa2d0a-296d-4ce8-8928-b0b6a4691e39", "address": "fa:16:3e:c3:4d:70", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6caa2d0a-29", "ovs_interfaceid": "6caa2d0a-296d-4ce8-8928-b0b6a4691e39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1090.637457] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:4d:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6caa2d0a-296d-4ce8-8928-b0b6a4691e39', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1090.645260] env[61986]: DEBUG oslo.service.loopingcall [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.646387] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1090.646387] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bbf14ae1-1d60-4579-a93d-517c485cb64a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.663026] env[61986]: DEBUG nova.compute.utils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1090.663761] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1090.663934] env[61986]: DEBUG nova.network.neutron [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1090.673276] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1090.673276] env[61986]: value = "task-1160450" [ 1090.673276] env[61986]: _type = "Task" [ 1090.673276] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.682616] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160450, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.708279] env[61986]: DEBUG nova.policy [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c804063142764cac8244fd3d6bd71e16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838620f936dc4489be8b99ef87bf37ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1090.767180] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.767485] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.767802] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.768012] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.768196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.770928] env[61986]: INFO nova.compute.manager [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Terminating instance [ 1090.772883] env[61986]: DEBUG nova.compute.manager [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.773106] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1090.773991] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d77696e-7fc4-4343-8e36-e9fa3c0ddcc0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.781633] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1090.781888] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-010a3811-49b3-49a8-ad0e-33d2f5a8e174 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.787249] env[61986]: DEBUG oslo_vmware.api [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1090.787249] env[61986]: value = "task-1160451" [ 1090.787249] env[61986]: _type = "Task" [ 1090.787249] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.795690] env[61986]: DEBUG oslo_vmware.api [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.924236] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160446, 'name': ReconfigVM_Task, 'duration_secs': 0.519807} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.924236] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-1568453a-0d54-4730-b64d-cb77ca01e291/volume-1568453a-0d54-4730-b64d-cb77ca01e291.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.928198] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe85a1a9-9bd3-40f9-a3ea-b74c270fd5d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.944620] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1090.944620] env[61986]: value = "task-1160452" [ 1090.944620] env[61986]: _type = "Task" [ 1090.944620] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.953452] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.002245] env[61986]: DEBUG nova.network.neutron [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Successfully created port: 5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1091.097110] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.097271] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.097420] env[61986]: DEBUG nova.network.neutron [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.098622] env[61986]: DEBUG oslo_vmware.api [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Task: {'id': task-1160449, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216413} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.102043] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.102043] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.102043] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.102043] env[61986]: INFO nova.compute.manager [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1091.102043] env[61986]: DEBUG oslo.service.loopingcall [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.102043] env[61986]: DEBUG nova.compute.manager [-] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.102043] env[61986]: DEBUG nova.network.neutron [-] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.167497] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1091.187172] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160450, 'name': CreateVM_Task, 'duration_secs': 0.338594} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.187172] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1091.187172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.187172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.187172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1091.189361] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad97a53a-cae7-4027-aa0d-8cd522280e11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.194078] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1091.194078] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52174a7e-299e-05ea-af15-deb7cfe26c0a" [ 1091.194078] env[61986]: _type = "Task" [ 1091.194078] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.201967] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52174a7e-299e-05ea-af15-deb7cfe26c0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.249603] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance af31ba6f-0a3e-4a1c-af5e-617e8843839f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.249755] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance e825bf09-7f6c-45db-b6cf-9f2dabb98677 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.249880] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 221e10a3-da31-410c-80f8-4bcc2c515710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.249998] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 18f87ccd-ca34-44f6-aa7d-d38397cc479b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.250126] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.250237] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 785dc1af-bec0-447b-90f5-2cb1f53bb0f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.250353] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 9ba24893-f54f-414f-9cb6-01cc84d273ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.250460] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a6094cc4-7d17-4858-abae-972425241c8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.250642] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1091.250815] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1091.299108] env[61986]: DEBUG oslo_vmware.api [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160451, 'name': PowerOffVM_Task, 'duration_secs': 0.231491} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.299401] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1091.299572] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1091.299824] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41c4f8ea-e9ac-4c9f-9c6a-50a4bf575358 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.363159] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c560581-b014-4793-b914-5e10a68bf86a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.366855] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1091.367100] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1091.367289] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleting the datastore file [datastore1] af31ba6f-0a3e-4a1c-af5e-617e8843839f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.367885] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38d25e8f-414c-481f-ad59-5acce1d42a77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.372539] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aa425d-290e-493f-a6b3-e250be87051f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.376371] env[61986]: DEBUG oslo_vmware.api [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1091.376371] env[61986]: value = "task-1160454" [ 1091.376371] env[61986]: _type = "Task" [ 1091.376371] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.405805] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f56dd9-178c-45ac-9b92-2540a6fa1ee7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.413207] env[61986]: DEBUG oslo_vmware.api [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.413207] env[61986]: DEBUG nova.network.neutron [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updated VIF entry in instance network info cache for port 12d8dd0f-1d73-4e30-bb9e-f4e074639096. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1091.413207] env[61986]: DEBUG nova.network.neutron [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [{"id": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "address": "fa:16:3e:b9:a5:6d", "network": {"id": "000a1615-7114-4435-9d68-be865094337f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-68524031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5342ea3a8f664440a608b16f1b7b44ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12d8dd0f-1d", "ovs_interfaceid": "12d8dd0f-1d73-4e30-bb9e-f4e074639096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.418343] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ddfcb1-a2b0-415e-90fe-0ad8bf727b03 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.433126] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.457877] env[61986]: DEBUG oslo_vmware.api [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160452, 'name': ReconfigVM_Task, 'duration_secs': 0.158607} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.458223] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252470', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'name': 'volume-1568453a-0d54-4730-b64d-cb77ca01e291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e825bf09-7f6c-45db-b6cf-9f2dabb98677', 'attached_at': '', 'detached_at': '', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'serial': '1568453a-0d54-4730-b64d-cb77ca01e291'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1091.628930] env[61986]: DEBUG nova.network.neutron [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1091.705432] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52174a7e-299e-05ea-af15-deb7cfe26c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.011026} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.705798] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.706085] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.706349] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.706502] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.706717] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.707015] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb5ffb37-3076-4c1f-b889-858036a8fb89 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.716149] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.716224] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1091.716899] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dd2e7ff-ce87-49ff-8115-e4d05fa093e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.722152] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1091.722152] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234abbe-04be-a29a-2e75-1c91fd89df3b" [ 1091.722152] env[61986]: _type = "Task" [ 1091.722152] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.729610] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234abbe-04be-a29a-2e75-1c91fd89df3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.771318] env[61986]: DEBUG nova.network.neutron [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updating instance_info_cache with network_info: [{"id": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "address": "fa:16:3e:46:85:45", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78ca1d3d-a9", "ovs_interfaceid": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.889773] env[61986]: DEBUG oslo_vmware.api [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19886} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.890202] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.890504] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.890798] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.891098] env[61986]: INFO nova.compute.manager [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1091.891459] env[61986]: DEBUG oslo.service.loopingcall [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.891756] env[61986]: DEBUG nova.compute.manager [-] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.891888] env[61986]: DEBUG nova.network.neutron [-] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.909701] env[61986]: DEBUG nova.network.neutron [-] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.915294] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Releasing lock "refresh_cache-18f87ccd-ca34-44f6-aa7d-d38397cc479b" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.915658] env[61986]: DEBUG nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Received event network-vif-plugged-6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1091.915994] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Acquiring lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.916368] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.916670] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.916966] env[61986]: DEBUG nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] No waiting events found dispatching network-vif-plugged-6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1091.917261] env[61986]: WARNING nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Received unexpected event network-vif-plugged-6caa2d0a-296d-4ce8-8928-b0b6a4691e39 for instance with vm_state building and task_state spawning. [ 1091.917531] env[61986]: DEBUG nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Received event network-changed-6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1091.917783] env[61986]: DEBUG nova.compute.manager [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Refreshing instance network info cache due to event network-changed-6caa2d0a-296d-4ce8-8928-b0b6a4691e39. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1091.918091] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Acquiring lock "refresh_cache-785dc1af-bec0-447b-90f5-2cb1f53bb0f0" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.918327] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Acquired lock "refresh_cache-785dc1af-bec0-447b-90f5-2cb1f53bb0f0" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.918576] env[61986]: DEBUG nova.network.neutron [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Refreshing network info cache for port 6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1091.939470] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.953645] env[61986]: DEBUG nova.compute.manager [req-bae30b35-37a1-4de6-a252-b6cb7fd1eec2 req-e6836fa6-c2a3-4daa-966d-5cbb666d8359 service nova] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Received event network-vif-deleted-12d8dd0f-1d73-4e30-bb9e-f4e074639096 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.177028] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1092.208515] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1092.208975] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1092.209170] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.209401] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1092.209541] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.209987] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1092.209987] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1092.210170] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1092.211035] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1092.211035] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1092.211035] env[61986]: DEBUG nova.virt.hardware [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1092.211645] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a3ac5d-7bee-48f9-a462-781848c3997a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.220773] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa57db4-b658-4f3d-8be3-55b03b4cc44c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.240426] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234abbe-04be-a29a-2e75-1c91fd89df3b, 'name': SearchDatastore_Task, 'duration_secs': 0.011985} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.241238] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6337f07c-8e92-4834-bbdb-33211d035b3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.246617] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1092.246617] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52260e6c-f7f1-a24f-b87d-5e2a97497ab9" [ 1092.246617] env[61986]: _type = "Task" [ 1092.246617] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.256640] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52260e6c-f7f1-a24f-b87d-5e2a97497ab9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.276246] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.276559] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Instance network_info: |[{"id": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "address": "fa:16:3e:46:85:45", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78ca1d3d-a9", "ovs_interfaceid": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.276966] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:85:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78ca1d3d-a907-4b97-8eec-a75b313ca796', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.284857] env[61986]: DEBUG oslo.service.loopingcall [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.285082] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1092.285308] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f02a0b3-f54e-4534-98d9-3fe0f4a31e06 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.305053] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.305053] env[61986]: value = "task-1160455" [ 1092.305053] env[61986]: _type = "Task" [ 1092.305053] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.312331] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160455, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.346772] env[61986]: DEBUG nova.compute.manager [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Received event network-changed-78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.346977] env[61986]: DEBUG nova.compute.manager [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Refreshing instance network info cache due to event network-changed-78ca1d3d-a907-4b97-8eec-a75b313ca796. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1092.347216] env[61986]: DEBUG oslo_concurrency.lockutils [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] Acquiring lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.347364] env[61986]: DEBUG oslo_concurrency.lockutils [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] Acquired lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.347529] env[61986]: DEBUG nova.network.neutron [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Refreshing network info cache for port 78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.412935] env[61986]: INFO nova.compute.manager [-] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Took 1.31 seconds to deallocate network for instance. [ 1092.443970] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1092.444349] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.228s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.495605] env[61986]: DEBUG nova.objects.instance [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'flavor' on Instance uuid e825bf09-7f6c-45db-b6cf-9f2dabb98677 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.752619] env[61986]: DEBUG nova.network.neutron [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Updated VIF entry in instance network info cache for port 6caa2d0a-296d-4ce8-8928-b0b6a4691e39. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1092.752961] env[61986]: DEBUG nova.network.neutron [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Updating instance_info_cache with network_info: [{"id": "6caa2d0a-296d-4ce8-8928-b0b6a4691e39", "address": "fa:16:3e:c3:4d:70", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6caa2d0a-29", "ovs_interfaceid": "6caa2d0a-296d-4ce8-8928-b0b6a4691e39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.757559] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52260e6c-f7f1-a24f-b87d-5e2a97497ab9, 'name': SearchDatastore_Task, 'duration_secs': 0.01047} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.757915] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.758173] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 785dc1af-bec0-447b-90f5-2cb1f53bb0f0/785dc1af-bec0-447b-90f5-2cb1f53bb0f0.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1092.758432] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c89786a5-59d7-457d-ba70-cd9d9e6c694f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.766039] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1092.766039] env[61986]: value = "task-1160456" [ 1092.766039] env[61986]: _type = "Task" [ 1092.766039] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.774272] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.815491] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160455, 'name': CreateVM_Task, 'duration_secs': 0.396266} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.815649] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1092.816326] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.816497] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.816815] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1092.817067] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60a5982b-7f09-4a16-879f-baaf5ca09255 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.821758] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1092.821758] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274b437-d089-8c1c-78b7-5da406f59949" [ 1092.821758] env[61986]: _type = "Task" [ 1092.821758] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.832261] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274b437-d089-8c1c-78b7-5da406f59949, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.839340] env[61986]: DEBUG nova.network.neutron [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Successfully updated port: 5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.840676] env[61986]: DEBUG nova.network.neutron [-] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.922936] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.923236] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.923480] env[61986]: DEBUG nova.objects.instance [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lazy-loading 'resources' on Instance uuid 18f87ccd-ca34-44f6-aa7d-d38397cc479b {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.001290] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f0f5968e-bcaa-41f3-9447-5407edef98d1 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.272s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.104276] env[61986]: DEBUG nova.network.neutron [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updated VIF entry in instance network info cache for port 78ca1d3d-a907-4b97-8eec-a75b313ca796. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.104608] env[61986]: DEBUG nova.network.neutron [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updating instance_info_cache with network_info: [{"id": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "address": "fa:16:3e:46:85:45", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78ca1d3d-a9", "ovs_interfaceid": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.259367] env[61986]: DEBUG oslo_concurrency.lockutils [req-af71ed41-f5a4-43ec-9416-f15584eabd12 req-96355803-a7b6-4d58-b8b0-699f1821ea8a service nova] Releasing lock "refresh_cache-785dc1af-bec0-447b-90f5-2cb1f53bb0f0" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.275522] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496288} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.275803] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 785dc1af-bec0-447b-90f5-2cb1f53bb0f0/785dc1af-bec0-447b-90f5-2cb1f53bb0f0.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1093.276034] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1093.276297] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2033cecc-065e-425e-a4ee-933a41fad1c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.284921] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1093.284921] env[61986]: value = "task-1160457" [ 1093.284921] env[61986]: _type = "Task" [ 1093.284921] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.292672] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160457, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.332636] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274b437-d089-8c1c-78b7-5da406f59949, 'name': SearchDatastore_Task, 'duration_secs': 0.011122} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.332919] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.333142] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.333378] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.333528] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.333707] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.334045] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d7a4c36-c63c-4e68-8e9b-b5dd53244b2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.342519] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.342519] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1093.343504] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-101ff3dd-ebe6-459d-8001-9643525eccd5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.345790] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.345892] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.346040] env[61986]: DEBUG nova.network.neutron [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1093.347227] env[61986]: INFO nova.compute.manager [-] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Took 1.46 seconds to deallocate network for instance. [ 1093.354289] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1093.354289] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52054e37-fa3b-ecf9-744f-db65b3b543ff" [ 1093.354289] env[61986]: _type = "Task" [ 1093.354289] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.364286] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52054e37-fa3b-ecf9-744f-db65b3b543ff, 'name': SearchDatastore_Task, 'duration_secs': 0.008853} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.365050] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5149890d-2cb1-44f0-9f85-5d345f843d34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.370248] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1093.370248] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e00714-5e88-e5bc-e0ed-d349216ee365" [ 1093.370248] env[61986]: _type = "Task" [ 1093.370248] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.378195] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e00714-5e88-e5bc-e0ed-d349216ee365, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.539726] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04feabf2-4a78-4748-9e27-11eb365df458 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.547573] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18316519-af49-4077-918b-57775f9a0ddf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.579053] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7f73ce-85c5-4950-ab0f-ab72981a3350 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.587035] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cf2dd5-25bc-45d5-b519-ca3589daea29 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.600338] env[61986]: DEBUG nova.compute.provider_tree [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.608611] env[61986]: DEBUG oslo_concurrency.lockutils [req-74c69da5-ee07-4273-b159-d70017906b48 req-7d2dbffc-2c49-4b65-8b2b-c69cc9bec62f service nova] Releasing lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.794902] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160457, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.052312} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.795232] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1093.795983] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a125875a-4a08-4adf-9e5b-f36cb0cb40b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.817596] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 785dc1af-bec0-447b-90f5-2cb1f53bb0f0/785dc1af-bec0-447b-90f5-2cb1f53bb0f0.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.818121] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e1a6e2c-09b3-414a-8ad7-fc2ef8eb937b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.838867] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1093.838867] env[61986]: value = "task-1160458" [ 1093.838867] env[61986]: _type = "Task" [ 1093.838867] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.846628] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160458, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.853594] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.880619] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e00714-5e88-e5bc-e0ed-d349216ee365, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.881494] env[61986]: DEBUG nova.network.neutron [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1093.883378] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.883595] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 9ba24893-f54f-414f-9cb6-01cc84d273ca/9ba24893-f54f-414f-9cb6-01cc84d273ca.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1093.883869] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aca30420-006f-4f8e-9e1d-226b2ffc2883 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.891605] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1093.891605] env[61986]: value = "task-1160459" [ 1093.891605] env[61986]: _type = "Task" [ 1093.891605] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.900488] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160459, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.032166] env[61986]: DEBUG nova.network.neutron [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [{"id": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "address": "fa:16:3e:22:d2:28", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5de45ae0-ff", "ovs_interfaceid": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.104682] env[61986]: DEBUG nova.scheduler.client.report [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.182217] env[61986]: DEBUG nova.compute.manager [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Stashing vm_state: active {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1094.350897] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160458, 'name': ReconfigVM_Task, 'duration_secs': 0.444054} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.351277] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 785dc1af-bec0-447b-90f5-2cb1f53bb0f0/785dc1af-bec0-447b-90f5-2cb1f53bb0f0.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.351985] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ec2c0b8-a5a3-41f1-9434-e7846942ce02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.359406] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1094.359406] env[61986]: value = "task-1160460" [ 1094.359406] env[61986]: _type = "Task" [ 1094.359406] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.369498] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160460, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.379660] env[61986]: DEBUG nova.compute.manager [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Received event network-vif-deleted-b4336d9c-969c-4d2b-830f-c988c73c8410 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.379660] env[61986]: DEBUG nova.compute.manager [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Received event network-vif-plugged-5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.379660] env[61986]: DEBUG oslo_concurrency.lockutils [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] Acquiring lock "a6094cc4-7d17-4858-abae-972425241c8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.379660] env[61986]: DEBUG oslo_concurrency.lockutils [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] Lock "a6094cc4-7d17-4858-abae-972425241c8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.379660] env[61986]: DEBUG oslo_concurrency.lockutils [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] Lock "a6094cc4-7d17-4858-abae-972425241c8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.379660] env[61986]: DEBUG nova.compute.manager [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] No waiting events found dispatching network-vif-plugged-5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1094.379660] env[61986]: WARNING nova.compute.manager [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Received unexpected event network-vif-plugged-5de45ae0-ff50-4fae-8941-3bd85428a59e for instance with vm_state building and task_state spawning. [ 1094.380322] env[61986]: DEBUG nova.compute.manager [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Received event network-changed-5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.380322] env[61986]: DEBUG nova.compute.manager [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Refreshing instance network info cache due to event network-changed-5de45ae0-ff50-4fae-8941-3bd85428a59e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1094.380322] env[61986]: DEBUG oslo_concurrency.lockutils [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] Acquiring lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.403460] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160459, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.445150] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.445429] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.445612] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1094.535183] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.535524] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Instance network_info: |[{"id": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "address": "fa:16:3e:22:d2:28", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5de45ae0-ff", "ovs_interfaceid": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1094.535839] env[61986]: DEBUG oslo_concurrency.lockutils [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] Acquired lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.536063] env[61986]: DEBUG nova.network.neutron [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Refreshing network info cache for port 5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.537530] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:d2:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5de45ae0-ff50-4fae-8941-3bd85428a59e', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.545319] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating folder: Project (838620f936dc4489be8b99ef87bf37ce). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1094.548162] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-261fc642-c0c5-4498-a3d3-8b2a05990934 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.562567] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created folder: Project (838620f936dc4489be8b99ef87bf37ce) in parent group-v252271. [ 1094.562758] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating folder: Instances. Parent ref: group-v252473. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1094.562990] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7267085-f1a6-4893-86df-0579a2a333f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.572672] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created folder: Instances in parent group-v252473. [ 1094.572897] env[61986]: DEBUG oslo.service.loopingcall [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.573114] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1094.573318] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc684109-10dd-4868-908b-77082ab0d06a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.594788] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.594788] env[61986]: value = "task-1160463" [ 1094.594788] env[61986]: _type = "Task" [ 1094.594788] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.602166] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160463, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.611082] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.612846] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.759s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.613449] env[61986]: DEBUG nova.objects.instance [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'resources' on Instance uuid af31ba6f-0a3e-4a1c-af5e-617e8843839f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.629427] env[61986]: INFO nova.scheduler.client.report [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Deleted allocations for instance 18f87ccd-ca34-44f6-aa7d-d38397cc479b [ 1094.707302] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.809596] env[61986]: DEBUG nova.network.neutron [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updated VIF entry in instance network info cache for port 5de45ae0-ff50-4fae-8941-3bd85428a59e. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1094.810030] env[61986]: DEBUG nova.network.neutron [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [{"id": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "address": "fa:16:3e:22:d2:28", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5de45ae0-ff", "ovs_interfaceid": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.869766] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160460, 'name': Rename_Task, 'duration_secs': 0.149649} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.870394] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1094.870394] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e77bd0e-6d83-490d-a138-d01c69f0dc73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.876751] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1094.876751] env[61986]: value = "task-1160464" [ 1094.876751] env[61986]: _type = "Task" [ 1094.876751] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.884178] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160464, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.900602] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160459, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51928} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.900957] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 9ba24893-f54f-414f-9cb6-01cc84d273ca/9ba24893-f54f-414f-9cb6-01cc84d273ca.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1094.901205] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1094.901440] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb1bd1ae-58a6-423b-af2c-35190e00554c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.907748] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1094.907748] env[61986]: value = "task-1160465" [ 1094.907748] env[61986]: _type = "Task" [ 1094.907748] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.916155] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.108024] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160463, 'name': CreateVM_Task, 'duration_secs': 0.333598} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.108024] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1095.108024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.108024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.108024] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1095.108024] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1c35732-b404-4806-b03a-f1ab36e2730c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.111115] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1095.111115] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e0a2f1-ee2b-3ed5-393d-6f710da5820d" [ 1095.111115] env[61986]: _type = "Task" [ 1095.111115] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.121064] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e0a2f1-ee2b-3ed5-393d-6f710da5820d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.139010] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac9cface-6eeb-436a-b35e-84d45bfb6732 tempest-ServerRescueTestJSONUnderV235-1877794561 tempest-ServerRescueTestJSONUnderV235-1877794561-project-member] Lock "18f87ccd-ca34-44f6-aa7d-d38397cc479b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.152s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.223307] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791e8f28-7de8-441b-a057-94c207473770 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.233535] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512b06c6-5708-433f-b13b-724a15d97345 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.265541] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0844248c-06fc-4ef8-91d3-55d32f820175 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.274323] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20ba947-df98-44b0-bda8-e10e8f4eac09 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.288243] env[61986]: DEBUG nova.compute.provider_tree [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.314116] env[61986]: DEBUG oslo_concurrency.lockutils [req-70f6a711-23a3-4a54-b262-3b7595b4bdd3 req-316a6f5f-c1c8-4a17-993f-449924133e28 service nova] Releasing lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.388052] env[61986]: DEBUG oslo_vmware.api [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160464, 'name': PowerOnVM_Task, 'duration_secs': 0.495271} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.388193] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1095.388314] env[61986]: INFO nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1095.388498] env[61986]: DEBUG nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.389345] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d85031-a321-4a96-9661-44c02d4a45e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.417224] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072572} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.417499] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1095.418335] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c60cd9-8860-4259-96f6-8eb2d579822e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.442147] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 9ba24893-f54f-414f-9cb6-01cc84d273ca/9ba24893-f54f-414f-9cb6-01cc84d273ca.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.443201] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55229d98-91bb-436f-918a-66337602308d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.467468] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1095.467468] env[61986]: value = "task-1160466" [ 1095.467468] env[61986]: _type = "Task" [ 1095.467468] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.478084] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160466, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.478881] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.479016] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.479165] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1095.622299] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e0a2f1-ee2b-3ed5-393d-6f710da5820d, 'name': SearchDatastore_Task, 'duration_secs': 0.019437} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.622624] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.622861] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.623108] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.623261] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.623444] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.623801] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-173f6a07-13f3-4b11-9600-a64f3c482f08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.634423] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.634620] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1095.635363] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9af74b99-350c-4e05-b872-edfb32a19f45 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.641865] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1095.641865] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b805ca-1b4e-07c0-b32d-46a4388254f2" [ 1095.641865] env[61986]: _type = "Task" [ 1095.641865] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.649789] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b805ca-1b4e-07c0-b32d-46a4388254f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.791586] env[61986]: DEBUG nova.scheduler.client.report [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.908419] env[61986]: INFO nova.compute.manager [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Took 12.53 seconds to build instance. [ 1095.978229] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160466, 'name': ReconfigVM_Task, 'duration_secs': 0.258859} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.978521] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 9ba24893-f54f-414f-9cb6-01cc84d273ca/9ba24893-f54f-414f-9cb6-01cc84d273ca.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.979203] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce41df0d-573a-466b-9452-b404a9b623bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.987242] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1095.987242] env[61986]: value = "task-1160467" [ 1095.987242] env[61986]: _type = "Task" [ 1095.987242] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.995446] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160467, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.004032] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1096.153516] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b805ca-1b4e-07c0-b32d-46a4388254f2, 'name': SearchDatastore_Task, 'duration_secs': 0.009785} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.154358] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77e2942f-4846-4fe7-a3c3-dfb2cdf599b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.161287] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1096.161287] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52124e73-51c4-9c98-3b88-72bfdbfc7e6c" [ 1096.161287] env[61986]: _type = "Task" [ 1096.161287] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.169888] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52124e73-51c4-9c98-3b88-72bfdbfc7e6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.297774] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.685s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.299864] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.593s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.320834] env[61986]: INFO nova.scheduler.client.report [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted allocations for instance af31ba6f-0a3e-4a1c-af5e-617e8843839f [ 1096.412283] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f8a7727c-7117-4416-af41-6050bf005813 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.039s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.498928] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160467, 'name': Rename_Task, 'duration_secs': 0.125397} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.499331] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1096.499622] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a08f9255-8c6b-4429-97a3-bcc7f5dde9e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.508826] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1096.508826] env[61986]: value = "task-1160468" [ 1096.508826] env[61986]: _type = "Task" [ 1096.508826] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.517894] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.603987] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.673490] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52124e73-51c4-9c98-3b88-72bfdbfc7e6c, 'name': SearchDatastore_Task, 'duration_secs': 0.027349} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.673889] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.674281] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] a6094cc4-7d17-4858-abae-972425241c8f/a6094cc4-7d17-4858-abae-972425241c8f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1096.674636] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ddc5e70-aeca-4915-b9f1-0d41f20cc366 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.683388] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1096.683388] env[61986]: value = "task-1160469" [ 1096.683388] env[61986]: _type = "Task" [ 1096.683388] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.695999] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160469, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.804790] env[61986]: INFO nova.compute.claims [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1096.828592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ece95600-1677-4989-8b66-0c455afc0ae3 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "af31ba6f-0a3e-4a1c-af5e-617e8843839f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.061s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.020748] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160468, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.106316] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-af31ba6f-0a3e-4a1c-af5e-617e8843839f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.106560] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1097.106774] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.106934] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.107124] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.107312] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.107462] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.196333] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160469, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.311129] env[61986]: INFO nova.compute.resource_tracker [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating resource usage from migration bdfc7468-370a-4792-b625-c9cb7c123f71 [ 1097.418521] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0d57ca-1299-45fd-bcd4-2f31733c3df2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.426360] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb33cf13-1e7c-4a59-9afc-157ed7c00887 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.459267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.459612] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.459825] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.460025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.460214] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.462784] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6b2c3f-b13a-445f-9a24-e8be8b3d20f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.465781] env[61986]: INFO nova.compute.manager [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Terminating instance [ 1097.467863] env[61986]: DEBUG nova.compute.manager [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1097.468067] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1097.469264] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f50ee4-ac81-44fc-89a2-57ee448f770c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.475533] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8679eb52-470f-4eca-8e3c-64ad9cecf408 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.481404] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1097.482035] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a153eef2-9d30-4c84-87e4-b40d202c356f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.493555] env[61986]: DEBUG nova.compute.provider_tree [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.500037] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1097.500037] env[61986]: value = "task-1160470" [ 1097.500037] env[61986]: _type = "Task" [ 1097.500037] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.508617] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.518575] env[61986]: DEBUG oslo_vmware.api [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160468, 'name': PowerOnVM_Task, 'duration_secs': 0.570503} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.518829] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1097.519049] env[61986]: INFO nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Took 7.60 seconds to spawn the instance on the hypervisor. [ 1097.519239] env[61986]: DEBUG nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.520047] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbf8161-765c-4067-bd99-4005478a98cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.694600] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160469, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613352} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.694881] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] a6094cc4-7d17-4858-abae-972425241c8f/a6094cc4-7d17-4858-abae-972425241c8f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1097.695121] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1097.695397] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6ef6801-4739-49ef-bf95-f68064830a4e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.703289] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1097.703289] env[61986]: value = "task-1160471" [ 1097.703289] env[61986]: _type = "Task" [ 1097.703289] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.712204] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.997300] env[61986]: DEBUG nova.scheduler.client.report [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.011677] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160470, 'name': PowerOffVM_Task, 'duration_secs': 0.311628} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.013249] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1098.013249] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1098.013249] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd81387f-f649-49e5-ad73-2a75098ab39e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.037639] env[61986]: INFO nova.compute.manager [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Took 12.48 seconds to build instance. [ 1098.098623] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1098.098859] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1098.099174] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore2] 785dc1af-bec0-447b-90f5-2cb1f53bb0f0 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.099481] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a008ddf-13a0-4496-b6af-417c96cf0370 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.108529] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1098.108529] env[61986]: value = "task-1160473" [ 1098.108529] env[61986]: _type = "Task" [ 1098.108529] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.117293] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160473, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.214284] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130737} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.214613] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1098.215464] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa23b1b-9e13-4a8a-9f26-61f75bf231e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.238189] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] a6094cc4-7d17-4858-abae-972425241c8f/a6094cc4-7d17-4858-abae-972425241c8f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.238519] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f378c051-a94f-4518-912b-59a577cb9307 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.259585] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1098.259585] env[61986]: value = "task-1160474" [ 1098.259585] env[61986]: _type = "Task" [ 1098.259585] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.268282] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160474, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.506644] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.207s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.506873] env[61986]: INFO nova.compute.manager [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Migrating [ 1098.540990] env[61986]: DEBUG oslo_concurrency.lockutils [None req-78e975dc-0cc0-4457-b0ef-2f772239191f tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.995s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.619879] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160473, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.774509] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.024030] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.024030] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.024030] env[61986]: DEBUG nova.network.neutron [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1099.119807] env[61986]: DEBUG oslo_vmware.api [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160473, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.526388} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.120157] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.120259] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1099.120437] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1099.120617] env[61986]: INFO nova.compute.manager [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1099.120886] env[61986]: DEBUG oslo.service.loopingcall [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.121105] env[61986]: DEBUG nova.compute.manager [-] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1099.121203] env[61986]: DEBUG nova.network.neutron [-] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1099.126930] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.126930] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.126930] env[61986]: INFO nova.compute.manager [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Shelving [ 1099.273683] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160474, 'name': ReconfigVM_Task, 'duration_secs': 0.819586} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.273683] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Reconfigured VM instance instance-00000067 to attach disk [datastore2] a6094cc4-7d17-4858-abae-972425241c8f/a6094cc4-7d17-4858-abae-972425241c8f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.273683] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a7b5a1f-1290-43a7-869e-1f292b7b97a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.280206] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1099.280206] env[61986]: value = "task-1160475" [ 1099.280206] env[61986]: _type = "Task" [ 1099.280206] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.288107] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160475, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.382778] env[61986]: DEBUG nova.compute.manager [req-81fa3a03-7ac1-4ca1-a5e5-a74d4cb866d6 req-8eb61891-14fd-4190-8077-54d4411d1ebd service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Received event network-vif-deleted-6caa2d0a-296d-4ce8-8928-b0b6a4691e39 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1099.382918] env[61986]: INFO nova.compute.manager [req-81fa3a03-7ac1-4ca1-a5e5-a74d4cb866d6 req-8eb61891-14fd-4190-8077-54d4411d1ebd service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Neutron deleted interface 6caa2d0a-296d-4ce8-8928-b0b6a4691e39; detaching it from the instance and deleting it from the info cache [ 1099.383132] env[61986]: DEBUG nova.network.neutron [req-81fa3a03-7ac1-4ca1-a5e5-a74d4cb866d6 req-8eb61891-14fd-4190-8077-54d4411d1ebd service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.571311] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.571544] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.633987] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1099.634771] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6ced54a-23e4-46f6-ba95-05928d9e493a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.641714] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1099.641714] env[61986]: value = "task-1160476" [ 1099.641714] env[61986]: _type = "Task" [ 1099.641714] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.652057] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.769662] env[61986]: DEBUG nova.network.neutron [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.790964] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160475, 'name': Rename_Task, 'duration_secs': 0.148939} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.791295] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1099.791625] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d811620f-b592-41f2-93e2-2512ad55cd21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.798409] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1099.798409] env[61986]: value = "task-1160477" [ 1099.798409] env[61986]: _type = "Task" [ 1099.798409] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.806844] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.858392] env[61986]: DEBUG nova.network.neutron [-] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.886098] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ede64c4b-3c6b-491e-8fe2-6dfd74d75a5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.895933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b198622-c0d0-4c3f-a493-6d9283043302 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.924025] env[61986]: DEBUG nova.compute.manager [req-81fa3a03-7ac1-4ca1-a5e5-a74d4cb866d6 req-8eb61891-14fd-4190-8077-54d4411d1ebd service nova] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Detach interface failed, port_id=6caa2d0a-296d-4ce8-8928-b0b6a4691e39, reason: Instance 785dc1af-bec0-447b-90f5-2cb1f53bb0f0 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1100.074205] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1100.151639] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160476, 'name': PowerOffVM_Task, 'duration_secs': 0.301868} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.152155] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1100.152611] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc359cc-cab3-4808-ad67-ab4f929f83df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.169773] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df531e0e-cb58-44ee-a89f-4e62da3913e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.273031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.309573] env[61986]: DEBUG oslo_vmware.api [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160477, 'name': PowerOnVM_Task, 'duration_secs': 0.486001} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.309816] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1100.310012] env[61986]: INFO nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1100.310213] env[61986]: DEBUG nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.310985] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec7a9d7-8044-4544-9d22-339c3fdab8e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.362008] env[61986]: INFO nova.compute.manager [-] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Took 1.24 seconds to deallocate network for instance. [ 1100.599958] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.600258] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.602049] env[61986]: INFO nova.compute.claims [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.679997] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1100.680338] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-04c80172-ff56-40a0-ba13-e141acc1e53a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.687858] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1100.687858] env[61986]: value = "task-1160478" [ 1100.687858] env[61986]: _type = "Task" [ 1100.687858] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.695756] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160478, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.830190] env[61986]: INFO nova.compute.manager [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Took 13.51 seconds to build instance. [ 1100.871017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.954093] env[61986]: DEBUG nova.compute.manager [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Received event network-changed-5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1100.954276] env[61986]: DEBUG nova.compute.manager [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Refreshing instance network info cache due to event network-changed-5de45ae0-ff50-4fae-8941-3bd85428a59e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1100.954496] env[61986]: DEBUG oslo_concurrency.lockutils [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] Acquiring lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.954646] env[61986]: DEBUG oslo_concurrency.lockutils [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] Acquired lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.954810] env[61986]: DEBUG nova.network.neutron [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Refreshing network info cache for port 5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1101.198262] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160478, 'name': CreateSnapshot_Task, 'duration_secs': 0.431935} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.198262] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1101.199503] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d902904-07eb-41a2-b357-bd34dd2d5e22 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.332988] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9733ac27-9d1f-43df-9742-578dc201e23e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "a6094cc4-7d17-4858-abae-972425241c8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.024s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.682367] env[61986]: DEBUG nova.network.neutron [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updated VIF entry in instance network info cache for port 5de45ae0-ff50-4fae-8941-3bd85428a59e. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1101.682723] env[61986]: DEBUG nova.network.neutron [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [{"id": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "address": "fa:16:3e:22:d2:28", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5de45ae0-ff", "ovs_interfaceid": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.717576] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1101.718872] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-580f493f-ee14-4d0c-bb95-1b234d44d2d4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.722215] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cf9a96-abda-4a98-99ba-0fbd062bbf7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.729530] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a9e9be-2165-4a46-9e63-2c3ba1ffad29 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.733299] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1101.733299] env[61986]: value = "task-1160479" [ 1101.733299] env[61986]: _type = "Task" [ 1101.733299] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.761464] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b939447-4647-4ad7-8eac-d6ad56910aeb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.766324] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160479, 'name': CloneVM_Task} progress is 12%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.770875] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f827931-dcd9-41e9-856a-9d5173c578c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.786417] env[61986]: DEBUG nova.compute.provider_tree [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.788476] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cbbb25-73e5-482f-8fbf-80206c19507d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.809296] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.185504] env[61986]: DEBUG oslo_concurrency.lockutils [req-f73c97b2-bc66-4d69-b64c-23503474c025 req-0eb0c2d4-a46a-4991-ac15-337540080b48 service nova] Releasing lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.243464] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160479, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.292759] env[61986]: DEBUG nova.scheduler.client.report [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.314578] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1102.314879] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-256e92b8-b61e-47f9-860e-fc7cf5aa917a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.322781] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1102.322781] env[61986]: value = "task-1160480" [ 1102.322781] env[61986]: _type = "Task" [ 1102.322781] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.331122] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.746487] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160479, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.797689] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.798306] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1102.801147] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.930s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.801371] env[61986]: DEBUG nova.objects.instance [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid 785dc1af-bec0-447b-90f5-2cb1f53bb0f0 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.833013] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160480, 'name': PowerOffVM_Task, 'duration_secs': 0.183341} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.833317] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1102.833512] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1103.245736] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160479, 'name': CloneVM_Task, 'duration_secs': 1.119378} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.247062] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Created linked-clone VM from snapshot [ 1103.247248] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7afee6-b35d-4dbd-8f26-45fc1960f38e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.255471] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Uploading image 45fcc477-9627-4bee-b93e-b423469cc638 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1103.280371] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1103.280371] env[61986]: value = "vm-252477" [ 1103.280371] env[61986]: _type = "VirtualMachine" [ 1103.280371] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1103.280648] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d787ed08-0891-42f4-8898-62e2bba5c9fd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.287414] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lease: (returnval){ [ 1103.287414] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336f8a-fe61-fc1d-6d36-d85c8b82d7a4" [ 1103.287414] env[61986]: _type = "HttpNfcLease" [ 1103.287414] env[61986]: } obtained for exporting VM: (result){ [ 1103.287414] env[61986]: value = "vm-252477" [ 1103.287414] env[61986]: _type = "VirtualMachine" [ 1103.287414] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1103.287727] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the lease: (returnval){ [ 1103.287727] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336f8a-fe61-fc1d-6d36-d85c8b82d7a4" [ 1103.287727] env[61986]: _type = "HttpNfcLease" [ 1103.287727] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1103.293492] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1103.293492] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336f8a-fe61-fc1d-6d36-d85c8b82d7a4" [ 1103.293492] env[61986]: _type = "HttpNfcLease" [ 1103.293492] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1103.303874] env[61986]: DEBUG nova.compute.utils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1103.307649] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1103.307823] env[61986]: DEBUG nova.network.neutron [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1103.339749] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.340019] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.340202] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.340392] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.340541] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.340691] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.340915] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.341092] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.341266] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.341430] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.341601] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.350266] env[61986]: DEBUG nova.policy [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8562656b1e46628059ea24f9e16b26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0db18c866dc4ffb8bcb050f8ec6021d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1103.351597] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fca3f5f-3322-4739-ae70-852a1adb2586 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.368184] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1103.368184] env[61986]: value = "task-1160482" [ 1103.368184] env[61986]: _type = "Task" [ 1103.368184] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.377559] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160482, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.434173] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07830684-4168-4ee6-964d-66da4ae188cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.441547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48072224-75b0-496b-8cbc-7f2e960319f0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.473087] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62dc1d8-819e-4351-8fc4-f9a6c2b5a667 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.480068] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e760ca-a466-4cac-ba04-13eafa618559 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.493741] env[61986]: DEBUG nova.compute.provider_tree [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.637741] env[61986]: DEBUG nova.network.neutron [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Successfully created port: dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1103.795809] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1103.795809] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336f8a-fe61-fc1d-6d36-d85c8b82d7a4" [ 1103.795809] env[61986]: _type = "HttpNfcLease" [ 1103.795809] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1103.796135] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1103.796135] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336f8a-fe61-fc1d-6d36-d85c8b82d7a4" [ 1103.796135] env[61986]: _type = "HttpNfcLease" [ 1103.796135] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1103.796885] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8b13f4-2fff-4d01-b215-0e22c466c475 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.804190] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5296761d-39a3-6175-5931-386c06854c7a/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1103.804374] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5296761d-39a3-6175-5931-386c06854c7a/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1103.864301] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1103.878523] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160482, 'name': ReconfigVM_Task, 'duration_secs': 0.496773} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.878854] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1103.898494] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-be4abeb3-d823-498f-8989-e848d283b091 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.997019] env[61986]: DEBUG nova.scheduler.client.report [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.385293] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.385293] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.385593] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.385908] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.386098] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.386287] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.386507] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.386712] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.386933] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.387140] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.387355] env[61986]: DEBUG nova.virt.hardware [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.395504] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1104.395863] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6bdac18-6c81-4ec5-9b88-ceb4060e2788 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.416968] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1104.416968] env[61986]: value = "task-1160483" [ 1104.416968] env[61986]: _type = "Task" [ 1104.416968] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.426564] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.501729] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.530236] env[61986]: INFO nova.scheduler.client.report [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance 785dc1af-bec0-447b-90f5-2cb1f53bb0f0 [ 1104.876147] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1104.899945] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.900483] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.900483] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.901535] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.901535] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.901535] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.901672] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.901815] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.901992] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.902186] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.903042] env[61986]: DEBUG nova.virt.hardware [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.903854] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f083dd-1f50-4182-a8d7-1ff4a5829516 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.913103] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa8851c-1bfd-4077-bf63-abcedcb748dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.927821] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160483, 'name': ReconfigVM_Task, 'duration_secs': 0.210595} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.936663] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1104.937824] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a885d8-6b25-4d62-b1ae-4d5a0c43f31d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.964015] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.964416] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-817f7bfa-7559-457d-8def-dc454eebdea4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.983618] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1104.983618] env[61986]: value = "task-1160484" [ 1104.983618] env[61986]: _type = "Task" [ 1104.983618] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.993200] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160484, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.038739] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3d54ad8e-3c52-4039-bdf8-62fcc3f19f9c tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "785dc1af-bec0-447b-90f5-2cb1f53bb0f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.579s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.044223] env[61986]: DEBUG nova.compute.manager [req-bff557a1-dbf1-48c5-8e04-f3755133e53c req-a1af42af-0e33-42d8-89ad-3ecbc900d911 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Received event network-vif-plugged-dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.044223] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff557a1-dbf1-48c5-8e04-f3755133e53c req-a1af42af-0e33-42d8-89ad-3ecbc900d911 service nova] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.044223] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff557a1-dbf1-48c5-8e04-f3755133e53c req-a1af42af-0e33-42d8-89ad-3ecbc900d911 service nova] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.044223] env[61986]: DEBUG oslo_concurrency.lockutils [req-bff557a1-dbf1-48c5-8e04-f3755133e53c req-a1af42af-0e33-42d8-89ad-3ecbc900d911 service nova] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.044223] env[61986]: DEBUG nova.compute.manager [req-bff557a1-dbf1-48c5-8e04-f3755133e53c req-a1af42af-0e33-42d8-89ad-3ecbc900d911 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] No waiting events found dispatching network-vif-plugged-dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1105.044223] env[61986]: WARNING nova.compute.manager [req-bff557a1-dbf1-48c5-8e04-f3755133e53c req-a1af42af-0e33-42d8-89ad-3ecbc900d911 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Received unexpected event network-vif-plugged-dba5f768-392c-41ae-9284-b299120172b9 for instance with vm_state building and task_state spawning. [ 1105.133634] env[61986]: DEBUG nova.network.neutron [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Successfully updated port: dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1105.495076] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160484, 'name': ReconfigVM_Task, 'duration_secs': 0.483535} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.495076] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to attach disk [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.495076] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 50 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1105.642433] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.642526] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.642712] env[61986]: DEBUG nova.network.neutron [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.782965] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.782965] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.001746] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426d02fe-fede-4b9c-ba1c-8f89778689e8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.025175] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fe05fc-c8a0-4ad5-aaa6-8c17b034b0f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.045903] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 67 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1106.174468] env[61986]: DEBUG nova.network.neutron [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1106.287096] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1106.325181] env[61986]: DEBUG nova.network.neutron [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updating instance_info_cache with network_info: [{"id": "dba5f768-392c-41ae-9284-b299120172b9", "address": "fa:16:3e:1d:7c:01", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba5f768-39", "ovs_interfaceid": "dba5f768-392c-41ae-9284-b299120172b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.814500] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.814809] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.816468] env[61986]: INFO nova.compute.claims [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1106.828126] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.828427] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Instance network_info: |[{"id": "dba5f768-392c-41ae-9284-b299120172b9", "address": "fa:16:3e:1d:7c:01", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba5f768-39", "ovs_interfaceid": "dba5f768-392c-41ae-9284-b299120172b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1106.828832] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:7c:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dba5f768-392c-41ae-9284-b299120172b9', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1106.837312] env[61986]: DEBUG oslo.service.loopingcall [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.839151] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1106.839418] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dd9daad-a13c-43f4-903c-06c1d974a9a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.862284] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1106.862284] env[61986]: value = "task-1160485" [ 1106.862284] env[61986]: _type = "Task" [ 1106.862284] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.870498] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160485, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.068510] env[61986]: DEBUG nova.compute.manager [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Received event network-changed-dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1107.068760] env[61986]: DEBUG nova.compute.manager [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Refreshing instance network info cache due to event network-changed-dba5f768-392c-41ae-9284-b299120172b9. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1107.068998] env[61986]: DEBUG oslo_concurrency.lockutils [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] Acquiring lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.069169] env[61986]: DEBUG oslo_concurrency.lockutils [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] Acquired lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.069334] env[61986]: DEBUG nova.network.neutron [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Refreshing network info cache for port dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1107.373247] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160485, 'name': CreateVM_Task, 'duration_secs': 0.414648} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.373448] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1107.374199] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.374391] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.374726] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1107.375007] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04a6fc46-b8c2-4583-83ba-bbeab0546781 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.379826] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1107.379826] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709c77-d143-3ee8-432a-10b3571de030" [ 1107.379826] env[61986]: _type = "Task" [ 1107.379826] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.389465] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709c77-d143-3ee8-432a-10b3571de030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.714446] env[61986]: DEBUG nova.network.neutron [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Port 1a9aae37-5d10-441b-a606-1d2451e00510 binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1107.789069] env[61986]: DEBUG nova.network.neutron [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updated VIF entry in instance network info cache for port dba5f768-392c-41ae-9284-b299120172b9. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1107.789655] env[61986]: DEBUG nova.network.neutron [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updating instance_info_cache with network_info: [{"id": "dba5f768-392c-41ae-9284-b299120172b9", "address": "fa:16:3e:1d:7c:01", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba5f768-39", "ovs_interfaceid": "dba5f768-392c-41ae-9284-b299120172b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.891058] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52709c77-d143-3ee8-432a-10b3571de030, 'name': SearchDatastore_Task, 'duration_secs': 0.009435} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.893592] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.893836] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1107.894085] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.894239] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.894421] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1107.894854] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a13356e1-f67e-409c-b6f4-3652cdc520bc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.903148] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1107.903357] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1107.906252] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f3808ca-2c9a-41bc-829c-c010ed225e9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.911573] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1107.911573] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528ec425-04cf-37f6-6aeb-431f28db7810" [ 1107.911573] env[61986]: _type = "Task" [ 1107.911573] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.920307] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528ec425-04cf-37f6-6aeb-431f28db7810, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.946400] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c39f1d-571f-4cb0-be93-1e88394f464b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.954056] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b053e6-947a-482a-a176-43b5739eef74 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.984309] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3a9901-6716-4d54-9724-807ad1af24dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.992439] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567af044-a445-4c81-aa76-aaaef8005670 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.005866] env[61986]: DEBUG nova.compute.provider_tree [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.292197] env[61986]: DEBUG oslo_concurrency.lockutils [req-5445344d-418c-462b-8fed-9a681932adba req-e236ad4f-c3af-4985-8eda-f3515fbb01f7 service nova] Releasing lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.422337] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528ec425-04cf-37f6-6aeb-431f28db7810, 'name': SearchDatastore_Task, 'duration_secs': 0.008941} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.423274] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b9b31c8-cc47-47ae-b49d-858b471e075d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.428331] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1108.428331] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521f287a-cf55-54d1-6f4e-75ae91f169f3" [ 1108.428331] env[61986]: _type = "Task" [ 1108.428331] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.435860] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521f287a-cf55-54d1-6f4e-75ae91f169f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.509595] env[61986]: DEBUG nova.scheduler.client.report [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1108.738767] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.739079] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.739196] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.938348] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521f287a-cf55-54d1-6f4e-75ae91f169f3, 'name': SearchDatastore_Task, 'duration_secs': 0.008744} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.938719] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.939066] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 39ea2c99-8b36-457e-96ef-6310223b0d7e/39ea2c99-8b36-457e-96ef-6310223b0d7e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1108.939144] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c98fdf5-41c0-4905-9b28-87e823b0d367 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.945279] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1108.945279] env[61986]: value = "task-1160486" [ 1108.945279] env[61986]: _type = "Task" [ 1108.945279] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.953182] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.016119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.201s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.016820] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1109.454913] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44052} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.455187] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 39ea2c99-8b36-457e-96ef-6310223b0d7e/39ea2c99-8b36-457e-96ef-6310223b0d7e.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1109.455444] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1109.455708] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-021b33a7-0c07-43d1-ad72-0368a0a4b6ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.461724] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1109.461724] env[61986]: value = "task-1160487" [ 1109.461724] env[61986]: _type = "Task" [ 1109.461724] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.469572] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160487, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.523223] env[61986]: DEBUG nova.compute.utils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1109.524937] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1109.524937] env[61986]: DEBUG nova.network.neutron [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1109.564836] env[61986]: DEBUG nova.policy [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1109.935655] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.935894] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.936105] env[61986]: DEBUG nova.network.neutron [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1109.972624] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160487, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094957} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.972984] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1109.973817] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7cb8ec-af99-4c73-90ed-51a65cbf3ac0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.002071] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 39ea2c99-8b36-457e-96ef-6310223b0d7e/39ea2c99-8b36-457e-96ef-6310223b0d7e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1110.002435] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5f18481-63ab-4bc7-844b-836436ed49ec {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.023680] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1110.023680] env[61986]: value = "task-1160488" [ 1110.023680] env[61986]: _type = "Task" [ 1110.023680] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.028734] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1110.038048] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160488, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.070258] env[61986]: DEBUG nova.network.neutron [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Successfully created port: 1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1110.539507] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160488, 'name': ReconfigVM_Task, 'duration_secs': 0.334745} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.540082] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 39ea2c99-8b36-457e-96ef-6310223b0d7e/39ea2c99-8b36-457e-96ef-6310223b0d7e.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.540736] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5bccd026-1cf4-469c-abe7-be6011409d27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.548364] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1110.548364] env[61986]: value = "task-1160489" [ 1110.548364] env[61986]: _type = "Task" [ 1110.548364] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.561028] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160489, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.645282] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5296761d-39a3-6175-5931-386c06854c7a/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1110.646327] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc843480-e669-4864-b971-9bc5f4743d17 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.652719] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5296761d-39a3-6175-5931-386c06854c7a/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1110.652898] env[61986]: ERROR oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5296761d-39a3-6175-5931-386c06854c7a/disk-0.vmdk due to incomplete transfer. [ 1110.653146] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-75eeba73-909a-42d7-a95a-4bdeea322f00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.659774] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5296761d-39a3-6175-5931-386c06854c7a/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1110.659974] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Uploaded image 45fcc477-9627-4bee-b93e-b423469cc638 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1110.662447] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1110.664821] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-24d4649f-2092-4800-b9e2-a83623779fff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.670746] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1110.670746] env[61986]: value = "task-1160490" [ 1110.670746] env[61986]: _type = "Task" [ 1110.670746] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.678981] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160490, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.700318] env[61986]: DEBUG nova.network.neutron [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.042269] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1111.057542] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160489, 'name': Rename_Task, 'duration_secs': 0.138552} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.059465] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1111.059975] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a4f6fb8-4217-4e38-a232-11751b1056f1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.062119] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.062384] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.066440] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1111.066668] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1111.066829] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1111.067028] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1111.067188] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1111.067345] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1111.067553] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1111.067713] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1111.067885] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1111.068065] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1111.068243] env[61986]: DEBUG nova.virt.hardware [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.068992] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b926d1-0e13-41c8-962b-9639fd5bb536 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.073081] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1111.073081] env[61986]: value = "task-1160491" [ 1111.073081] env[61986]: _type = "Task" [ 1111.073081] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.079040] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44334523-6cd9-40b4-86f6-337c58deb9a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.087234] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160491, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.181887] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160490, 'name': Destroy_Task, 'duration_secs': 0.275215} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.182231] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Destroyed the VM [ 1111.182474] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1111.182728] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bf158f76-6b2f-4cee-87c0-2b3fd7663b71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.189439] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1111.189439] env[61986]: value = "task-1160492" [ 1111.189439] env[61986]: _type = "Task" [ 1111.189439] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.197204] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160492, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.202844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.502988] env[61986]: DEBUG nova.compute.manager [req-c56e8c9e-94c9-4c90-8be7-046a17d667d5 req-447e84c9-3b60-4e39-aac3-b97e43aad2a6 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Received event network-vif-plugged-1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.503205] env[61986]: DEBUG oslo_concurrency.lockutils [req-c56e8c9e-94c9-4c90-8be7-046a17d667d5 req-447e84c9-3b60-4e39-aac3-b97e43aad2a6 service nova] Acquiring lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.503410] env[61986]: DEBUG oslo_concurrency.lockutils [req-c56e8c9e-94c9-4c90-8be7-046a17d667d5 req-447e84c9-3b60-4e39-aac3-b97e43aad2a6 service nova] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.503574] env[61986]: DEBUG oslo_concurrency.lockutils [req-c56e8c9e-94c9-4c90-8be7-046a17d667d5 req-447e84c9-3b60-4e39-aac3-b97e43aad2a6 service nova] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.503828] env[61986]: DEBUG nova.compute.manager [req-c56e8c9e-94c9-4c90-8be7-046a17d667d5 req-447e84c9-3b60-4e39-aac3-b97e43aad2a6 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] No waiting events found dispatching network-vif-plugged-1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1111.504139] env[61986]: WARNING nova.compute.manager [req-c56e8c9e-94c9-4c90-8be7-046a17d667d5 req-447e84c9-3b60-4e39-aac3-b97e43aad2a6 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Received unexpected event network-vif-plugged-1146e055-1ec5-4dff-9b99-ec02b64b7cfb for instance with vm_state building and task_state spawning. [ 1111.565279] env[61986]: INFO nova.compute.manager [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Detaching volume 3f7d5322-8141-4aa7-813d-2ae4f033983e [ 1111.582773] env[61986]: DEBUG oslo_vmware.api [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160491, 'name': PowerOnVM_Task, 'duration_secs': 0.433227} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.583422] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1111.583422] env[61986]: INFO nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Took 6.71 seconds to spawn the instance on the hypervisor. [ 1111.583593] env[61986]: DEBUG nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1111.586112] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06438f0-dfc7-4dc1-8d08-787085c492c1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.599530] env[61986]: DEBUG nova.network.neutron [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Successfully updated port: 1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1111.610885] env[61986]: INFO nova.virt.block_device [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Attempting to driver detach volume 3f7d5322-8141-4aa7-813d-2ae4f033983e from mountpoint /dev/sdb [ 1111.611320] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1111.612505] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252451', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'name': 'volume-3f7d5322-8141-4aa7-813d-2ae4f033983e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '221e10a3-da31-410c-80f8-4bcc2c515710', 'attached_at': '', 'detached_at': '', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'serial': '3f7d5322-8141-4aa7-813d-2ae4f033983e'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1111.612747] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6689eb-18ad-4393-87a5-5a0b0d46f0d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.635870] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22dbaef-23ba-4344-a067-cc738d51d00c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.644490] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f603573e-1f66-421b-936b-9eec8b4cef3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.665380] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4329a9d-2934-4f9d-8adb-9f64f9199203 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.682244] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] The volume has not been displaced from its original location: [datastore2] volume-3f7d5322-8141-4aa7-813d-2ae4f033983e/volume-3f7d5322-8141-4aa7-813d-2ae4f033983e.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1111.687524] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfiguring VM instance instance-00000052 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.687825] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b66ddefb-4570-42c7-9a9f-4ad41fe35dc7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.712744] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160492, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.713807] env[61986]: DEBUG oslo_vmware.api [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1111.713807] env[61986]: value = "task-1160493" [ 1111.713807] env[61986]: _type = "Task" [ 1111.713807] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.715050] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195f9865-e3a6-4cb5-b639-96208c1fb653 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.726457] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632bec67-97f7-429a-b36b-7cab7d6bf25e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.728655] env[61986]: DEBUG oslo_vmware.api [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160493, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.105963] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.106224] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.106308] env[61986]: DEBUG nova.network.neutron [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1112.110893] env[61986]: INFO nova.compute.manager [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Took 11.53 seconds to build instance. [ 1112.209836] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160492, 'name': RemoveSnapshot_Task, 'duration_secs': 0.599507} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.210126] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1112.210409] env[61986]: DEBUG nova.compute.manager [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1112.211193] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a48450d-b5df-4fc6-a069-e0582d0aad72 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.225079] env[61986]: DEBUG oslo_vmware.api [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160493, 'name': ReconfigVM_Task, 'duration_secs': 0.254194} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.225311] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Reconfigured VM instance instance-00000052 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1112.229790] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b3305ed-59cc-496f-ad2b-45f473ab1fad {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.244022] env[61986]: DEBUG oslo_vmware.api [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1112.244022] env[61986]: value = "task-1160494" [ 1112.244022] env[61986]: _type = "Task" [ 1112.244022] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.251922] env[61986]: DEBUG oslo_vmware.api [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160494, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.615987] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1e355f63-9fdd-4710-b6dd-b89818efcf6a tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.044s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.641251] env[61986]: DEBUG nova.network.neutron [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1112.725403] env[61986]: INFO nova.compute.manager [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Shelve offloading [ 1112.726531] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1112.726823] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-276e9f66-7c51-46a5-99c4-e69346963950 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.733851] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1112.733851] env[61986]: value = "task-1160495" [ 1112.733851] env[61986]: _type = "Task" [ 1112.733851] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.741334] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.752566] env[61986]: DEBUG oslo_vmware.api [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160494, 'name': ReconfigVM_Task, 'duration_secs': 0.135689} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.752848] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252451', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'name': 'volume-3f7d5322-8141-4aa7-813d-2ae4f033983e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '221e10a3-da31-410c-80f8-4bcc2c515710', 'attached_at': '', 'detached_at': '', 'volume_id': '3f7d5322-8141-4aa7-813d-2ae4f033983e', 'serial': '3f7d5322-8141-4aa7-813d-2ae4f033983e'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1112.780907] env[61986]: DEBUG nova.network.neutron [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Updating instance_info_cache with network_info: [{"id": "1146e055-1ec5-4dff-9b99-ec02b64b7cfb", "address": "fa:16:3e:ca:34:19", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1146e055-1e", "ovs_interfaceid": "1146e055-1ec5-4dff-9b99-ec02b64b7cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.822515] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85f16ce-ad41-4f0e-b988-23b4c1906606 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.845649] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfaf4529-3bf2-4605-a16d-99180c2da506 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.852164] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 83 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1113.027670] env[61986]: DEBUG nova.compute.manager [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Received event network-changed-dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1113.027872] env[61986]: DEBUG nova.compute.manager [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Refreshing instance network info cache due to event network-changed-dba5f768-392c-41ae-9284-b299120172b9. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1113.028153] env[61986]: DEBUG oslo_concurrency.lockutils [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] Acquiring lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.028310] env[61986]: DEBUG oslo_concurrency.lockutils [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] Acquired lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.028473] env[61986]: DEBUG nova.network.neutron [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Refreshing network info cache for port dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1113.245186] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1113.245546] env[61986]: DEBUG nova.compute.manager [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.246211] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beea278a-de9e-4dc6-824f-1275214197a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.252786] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.252956] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.253142] env[61986]: DEBUG nova.network.neutron [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1113.283300] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.283593] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Instance network_info: |[{"id": "1146e055-1ec5-4dff-9b99-ec02b64b7cfb", "address": "fa:16:3e:ca:34:19", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1146e055-1e", "ovs_interfaceid": "1146e055-1ec5-4dff-9b99-ec02b64b7cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1113.283981] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:34:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1146e055-1ec5-4dff-9b99-ec02b64b7cfb', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1113.291605] env[61986]: DEBUG oslo.service.loopingcall [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.292119] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1113.292355] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f3d0e36-6248-491a-84c5-fa315e9da15c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.307869] env[61986]: DEBUG nova.objects.instance [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'flavor' on Instance uuid 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.314824] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1113.314824] env[61986]: value = "task-1160496" [ 1113.314824] env[61986]: _type = "Task" [ 1113.314824] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.322894] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160496, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.358968] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1113.359302] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0bb64ec-300d-4e87-9a1b-2c06fe88a9c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.365593] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1113.365593] env[61986]: value = "task-1160497" [ 1113.365593] env[61986]: _type = "Task" [ 1113.365593] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.373706] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.530363] env[61986]: DEBUG nova.compute.manager [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Received event network-changed-1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1113.530644] env[61986]: DEBUG nova.compute.manager [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Refreshing instance network info cache due to event network-changed-1146e055-1ec5-4dff-9b99-ec02b64b7cfb. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1113.530791] env[61986]: DEBUG oslo_concurrency.lockutils [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] Acquiring lock "refresh_cache-fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.530985] env[61986]: DEBUG oslo_concurrency.lockutils [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] Acquired lock "refresh_cache-fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.531180] env[61986]: DEBUG nova.network.neutron [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Refreshing network info cache for port 1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1113.762745] env[61986]: DEBUG nova.network.neutron [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updated VIF entry in instance network info cache for port dba5f768-392c-41ae-9284-b299120172b9. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1113.763195] env[61986]: DEBUG nova.network.neutron [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updating instance_info_cache with network_info: [{"id": "dba5f768-392c-41ae-9284-b299120172b9", "address": "fa:16:3e:1d:7c:01", "network": {"id": "f07e886c-f90a-437c-ae18-986351ee118a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388912608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0db18c866dc4ffb8bcb050f8ec6021d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba5f768-39", "ovs_interfaceid": "dba5f768-392c-41ae-9284-b299120172b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.825483] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160496, 'name': CreateVM_Task, 'duration_secs': 0.334735} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.827722] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1113.828699] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.828699] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.828828] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1113.829803] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fefec231-e25f-4363-80d9-c34126a73233 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.834461] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1113.834461] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52563553-8373-f0d7-20a7-0158e333c613" [ 1113.834461] env[61986]: _type = "Task" [ 1113.834461] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.841642] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52563553-8373-f0d7-20a7-0158e333c613, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.873926] env[61986]: DEBUG oslo_vmware.api [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160497, 'name': PowerOnVM_Task, 'duration_secs': 0.39073} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.874211] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1113.874400] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2c7e1afb-49f2-4b06-a55e-c3936a3e832c tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance 'e825bf09-7f6c-45db-b6cf-9f2dabb98677' progress to 100 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1113.959642] env[61986]: DEBUG nova.network.neutron [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updating instance_info_cache with network_info: [{"id": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "address": "fa:16:3e:46:85:45", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78ca1d3d-a9", "ovs_interfaceid": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.228193] env[61986]: DEBUG nova.network.neutron [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Updated VIF entry in instance network info cache for port 1146e055-1ec5-4dff-9b99-ec02b64b7cfb. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1114.228617] env[61986]: DEBUG nova.network.neutron [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Updating instance_info_cache with network_info: [{"id": "1146e055-1ec5-4dff-9b99-ec02b64b7cfb", "address": "fa:16:3e:ca:34:19", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1146e055-1e", "ovs_interfaceid": "1146e055-1ec5-4dff-9b99-ec02b64b7cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.266070] env[61986]: DEBUG oslo_concurrency.lockutils [req-f99164a1-3974-40e3-b094-fbdfec33efa3 req-175b4329-f02e-4d53-850b-3e242bc27d4d service nova] Releasing lock "refresh_cache-39ea2c99-8b36-457e-96ef-6310223b0d7e" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.319782] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbb2a068-c581-47da-ad2d-dd43e1194163 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.257s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.345068] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52563553-8373-f0d7-20a7-0158e333c613, 'name': SearchDatastore_Task, 'duration_secs': 0.008845} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.345389] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.345625] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1114.345859] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.346016] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.346212] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.346484] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b12ad7b2-6cce-429b-9563-70c1ef5e05cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.354581] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.354759] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1114.355527] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9448e58-b23b-412f-a826-06efd276fd66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.360341] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1114.360341] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ba53aa-0bdc-9b01-4748-d91501ca8d12" [ 1114.360341] env[61986]: _type = "Task" [ 1114.360341] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.367743] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ba53aa-0bdc-9b01-4748-d91501ca8d12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.462722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.732308] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1114.732894] env[61986]: DEBUG oslo_concurrency.lockutils [req-14291b13-6f61-4ccb-8ff4-7c5c50104d7e req-2917329a-958b-4a0b-8651-c4aa0e85b587 service nova] Releasing lock "refresh_cache-fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.734036] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0fea62-7546-4cd5-bfd9-0c15ad313186 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.741452] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1114.741700] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf4fa868-ef85-4687-b685-ad11be9a7d8e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.802607] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1114.804358] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1114.804358] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleting the datastore file [datastore2] 9ba24893-f54f-414f-9cb6-01cc84d273ca {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.804358] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63e49af5-8b49-4108-b1ee-c3f05465f3fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.810272] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1114.810272] env[61986]: value = "task-1160499" [ 1114.810272] env[61986]: _type = "Task" [ 1114.810272] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.817643] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.871477] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ba53aa-0bdc-9b01-4748-d91501ca8d12, 'name': SearchDatastore_Task, 'duration_secs': 0.007706} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.872300] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f561c7f-068f-45ef-8dd4-59e05800e432 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.877243] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1114.877243] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52df133e-0af1-5e68-91f3-7b4244c3265f" [ 1114.877243] env[61986]: _type = "Task" [ 1114.877243] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.884845] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52df133e-0af1-5e68-91f3-7b4244c3265f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.321192] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.331750] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.332032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.332292] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.332489] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.332662] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.334839] env[61986]: INFO nova.compute.manager [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Terminating instance [ 1115.336610] env[61986]: DEBUG nova.compute.manager [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1115.336807] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1115.337581] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d7e73d-6368-446d-8faf-f65314f6e645 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.344538] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1115.344736] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-594eb110-45f6-444b-937b-55ffa58df786 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.351691] env[61986]: DEBUG oslo_vmware.api [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1115.351691] env[61986]: value = "task-1160500" [ 1115.351691] env[61986]: _type = "Task" [ 1115.351691] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.359794] env[61986]: DEBUG oslo_vmware.api [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.387642] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52df133e-0af1-5e68-91f3-7b4244c3265f, 'name': SearchDatastore_Task, 'duration_secs': 0.009424} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.387929] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.388212] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff/fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1115.388479] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88064095-d7a0-488f-b930-0d6461b69559 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.395167] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1115.395167] env[61986]: value = "task-1160501" [ 1115.395167] env[61986]: _type = "Task" [ 1115.395167] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.403250] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160501, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.559387] env[61986]: DEBUG nova.compute.manager [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Received event network-vif-unplugged-78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1115.559626] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] Acquiring lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.560381] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.560578] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.560759] env[61986]: DEBUG nova.compute.manager [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] No waiting events found dispatching network-vif-unplugged-78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1115.560996] env[61986]: WARNING nova.compute.manager [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Received unexpected event network-vif-unplugged-78ca1d3d-a907-4b97-8eec-a75b313ca796 for instance with vm_state shelved and task_state shelving_offloading. [ 1115.561207] env[61986]: DEBUG nova.compute.manager [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Received event network-changed-78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1115.561373] env[61986]: DEBUG nova.compute.manager [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Refreshing instance network info cache due to event network-changed-78ca1d3d-a907-4b97-8eec-a75b313ca796. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1115.561560] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] Acquiring lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.561700] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] Acquired lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.561867] env[61986]: DEBUG nova.network.neutron [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Refreshing network info cache for port 78ca1d3d-a907-4b97-8eec-a75b313ca796 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1115.824505] env[61986]: DEBUG oslo_vmware.api [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160499, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.589427} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.824976] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.825306] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1115.825632] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1115.852033] env[61986]: INFO nova.scheduler.client.report [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocations for instance 9ba24893-f54f-414f-9cb6-01cc84d273ca [ 1115.864970] env[61986]: DEBUG oslo_vmware.api [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160500, 'name': PowerOffVM_Task, 'duration_secs': 0.174421} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.865340] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1115.865561] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1115.865837] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8770d486-95d2-4fa2-9c0b-c6f8fc185ad7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.908440] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160501, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.926626] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1115.926896] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1115.927146] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleting the datastore file [datastore2] 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.927436] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ceed1b5-19cc-4fa6-ae0e-8dd0b54c3b35 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.933976] env[61986]: DEBUG oslo_vmware.api [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1115.933976] env[61986]: value = "task-1160503" [ 1115.933976] env[61986]: _type = "Task" [ 1115.933976] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.942600] env[61986]: DEBUG oslo_vmware.api [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.268341] env[61986]: DEBUG nova.network.neutron [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updated VIF entry in instance network info cache for port 78ca1d3d-a907-4b97-8eec-a75b313ca796. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1116.268760] env[61986]: DEBUG nova.network.neutron [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updating instance_info_cache with network_info: [{"id": "78ca1d3d-a907-4b97-8eec-a75b313ca796", "address": "fa:16:3e:46:85:45", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": null, "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap78ca1d3d-a9", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.360234] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.360537] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.360805] env[61986]: DEBUG nova.objects.instance [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'resources' on Instance uuid 9ba24893-f54f-414f-9cb6-01cc84d273ca {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.406440] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160501, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542209} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.406722] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff/fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1116.406957] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1116.407230] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5196405-6888-4362-ab5f-39ab91d724dd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.414121] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1116.414121] env[61986]: value = "task-1160504" [ 1116.414121] env[61986]: _type = "Task" [ 1116.414121] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.423986] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.443674] env[61986]: DEBUG oslo_vmware.api [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116283} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.443987] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.444160] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1116.444350] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1116.444544] env[61986]: INFO nova.compute.manager [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1116.444788] env[61986]: DEBUG oslo.service.loopingcall [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1116.444989] env[61986]: DEBUG nova.compute.manager [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1116.445105] env[61986]: DEBUG nova.network.neutron [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1116.736658] env[61986]: DEBUG nova.network.neutron [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Port 1a9aae37-5d10-441b-a606-1d2451e00510 binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1116.736931] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.737103] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.737275] env[61986]: DEBUG nova.network.neutron [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.772046] env[61986]: DEBUG oslo_concurrency.lockutils [req-5e558610-a163-4419-b93c-f3cae4824571 req-0c73cdb3-eb11-4947-97c4-3f42155063e8 service nova] Releasing lock "refresh_cache-9ba24893-f54f-414f-9cb6-01cc84d273ca" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.851048] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.863316] env[61986]: DEBUG nova.objects.instance [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'numa_topology' on Instance uuid 9ba24893-f54f-414f-9cb6-01cc84d273ca {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.923312] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160504, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062361} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.923628] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1116.924454] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5668134-d931-401f-b4f3-9fb038cc3b84 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.946369] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff/fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1116.946639] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba3b5b85-6060-4f49-95ab-8ec4db6e9690 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.967355] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1116.967355] env[61986]: value = "task-1160505" [ 1116.967355] env[61986]: _type = "Task" [ 1116.967355] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.975484] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.365516] env[61986]: DEBUG nova.objects.base [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Object Instance<9ba24893-f54f-414f-9cb6-01cc84d273ca> lazy-loaded attributes: resources,numa_topology {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1117.478825] env[61986]: DEBUG nova.network.neutron [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.483263] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.499616] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108ca714-030d-486f-b75b-3033eb76abee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.508624] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90d9955-6aab-4cf1-8246-64f55a1c34cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.540332] env[61986]: DEBUG nova.network.neutron [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.542441] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8364a637-c192-42b3-a3a8-3720c2633121 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.550519] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58d7888-c5e5-430f-8361-764cc885adfd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.565956] env[61986]: DEBUG nova.compute.provider_tree [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.586592] env[61986]: DEBUG nova.compute.manager [req-4531cd70-6615-4f44-9f0d-43673a2a1300 req-2fa7e498-2478-4ca8-aa9e-607228e7c3ab service nova] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Received event network-vif-deleted-a298ab4e-6b6f-4d7e-ae0b-7d16a86990c7 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1117.978407] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160505, 'name': ReconfigVM_Task, 'duration_secs': 0.819377} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.978713] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Reconfigured VM instance instance-00000069 to attach disk [datastore1] fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff/fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1117.979417] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-702d0d88-2737-46b3-aea1-c1da0f7a212f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.984601] env[61986]: INFO nova.compute.manager [-] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Took 1.54 seconds to deallocate network for instance. [ 1117.987282] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1117.987282] env[61986]: value = "task-1160506" [ 1117.987282] env[61986]: _type = "Task" [ 1117.987282] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.997250] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160506, 'name': Rename_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.047136] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.068624] env[61986]: DEBUG nova.scheduler.client.report [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.494881] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.500922] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160506, 'name': Rename_Task, 'duration_secs': 0.137176} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.501479] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1118.501758] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1f198b4-280e-4c19-9e4d-806ed77f47ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.509384] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1118.509384] env[61986]: value = "task-1160507" [ 1118.509384] env[61986]: _type = "Task" [ 1118.509384] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.517767] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.550751] env[61986]: DEBUG nova.compute.manager [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61986) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1118.573557] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.213s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.576713] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.082s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.576954] env[61986]: DEBUG nova.objects.instance [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'resources' on Instance uuid 221e10a3-da31-410c-80f8-4bcc2c515710 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.020553] env[61986]: DEBUG oslo_vmware.api [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160507, 'name': PowerOnVM_Task, 'duration_secs': 0.488496} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.020908] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1119.021155] env[61986]: INFO nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1119.021390] env[61986]: DEBUG nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1119.022272] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d021a78f-ed55-4b10-911c-76ebc08d88e0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.081431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b55dbf1f-2d64-44ad-be52-83c4b49a9507 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.955s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.086009] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.235s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.086275] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.086486] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.086660] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.088628] env[61986]: INFO nova.compute.manager [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Terminating instance [ 1119.090353] env[61986]: DEBUG nova.compute.manager [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1119.090546] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1119.090805] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d89ca93-fd5d-4b49-8a5d-57be00c6dfbd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.101231] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edea6900-e7ba-4fd6-9bcc-6d76bf0e0e30 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.129952] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ba24893-f54f-414f-9cb6-01cc84d273ca could not be found. [ 1119.130180] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1119.130362] env[61986]: INFO nova.compute.manager [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1119.130639] env[61986]: DEBUG oslo.service.loopingcall [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.133304] env[61986]: DEBUG nova.compute.manager [-] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.133393] env[61986]: DEBUG nova.network.neutron [-] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1119.206222] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a872495-67f2-478b-9a9b-15843bb5fa6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.214320] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a98d46-4df8-4d0b-b149-b68f78785a79 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.245651] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e5b489-8b7f-4b63-b143-073a1c8986c1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.253349] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a43bc51-8707-4095-8153-09a7212657d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.270088] env[61986]: DEBUG nova.compute.provider_tree [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.540840] env[61986]: INFO nova.compute.manager [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Took 12.75 seconds to build instance. [ 1119.670603] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.773185] env[61986]: DEBUG nova.scheduler.client.report [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.835866] env[61986]: DEBUG nova.network.neutron [-] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.043345] env[61986]: DEBUG oslo_concurrency.lockutils [None req-410f26f8-d255-48a1-8daa-e3f5984b2910 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.260s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.277720] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.280695] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.610s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.305279] env[61986]: INFO nova.scheduler.client.report [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted allocations for instance 221e10a3-da31-410c-80f8-4bcc2c515710 [ 1120.337867] env[61986]: INFO nova.compute.manager [-] [instance: 9ba24893-f54f-414f-9cb6-01cc84d273ca] Took 1.20 seconds to deallocate network for instance. [ 1120.521882] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.521882] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.522150] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.522351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.522520] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.525548] env[61986]: INFO nova.compute.manager [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Terminating instance [ 1120.527405] env[61986]: DEBUG nova.compute.manager [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1120.527699] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1120.528475] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86490bac-ac3d-4234-a986-6cc87d00ea15 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.536253] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1120.536514] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98e3cd76-7bad-4099-b75c-459b3af0ace6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.543921] env[61986]: DEBUG oslo_vmware.api [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1120.543921] env[61986]: value = "task-1160508" [ 1120.543921] env[61986]: _type = "Task" [ 1120.543921] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.553990] env[61986]: DEBUG oslo_vmware.api [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.784160] env[61986]: DEBUG nova.objects.instance [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'migration_context' on Instance uuid e825bf09-7f6c-45db-b6cf-9f2dabb98677 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.815281] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac917ed4-f3ed-45c3-80ea-85603285d51d tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "221e10a3-da31-410c-80f8-4bcc2c515710" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.483s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.056714] env[61986]: DEBUG oslo_vmware.api [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160508, 'name': PowerOffVM_Task, 'duration_secs': 0.237195} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.056714] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1121.056714] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1121.056714] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11168381-ac4d-4659-8bb3-9d0088fc13cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.113376] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1121.113602] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1121.113788] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore1] fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1121.114081] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-805069ce-972a-4796-aa33-d87e4e136b12 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.120049] env[61986]: DEBUG oslo_vmware.api [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1121.120049] env[61986]: value = "task-1160510" [ 1121.120049] env[61986]: _type = "Task" [ 1121.120049] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.127794] env[61986]: DEBUG oslo_vmware.api [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160510, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.367916] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a858d5dc-e0f1-4bbb-bfbc-dee7a3144b16 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "9ba24893-f54f-414f-9cb6-01cc84d273ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.282s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.374464] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e39f95b-e7b9-434e-9fa9-9fd99fdb084e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.383968] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b74eb05-ffd7-482f-9beb-76aa6c7968a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.416462] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1e748a-b9c4-42a1-98e1-222152f3d733 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.427536] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0635a0-6946-4974-bdb0-4458df244dac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.442813] env[61986]: DEBUG nova.compute.provider_tree [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.631551] env[61986]: DEBUG oslo_vmware.api [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160510, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145815} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.631899] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.632333] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1121.632333] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1121.632470] env[61986]: INFO nova.compute.manager [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1121.632682] env[61986]: DEBUG oslo.service.loopingcall [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1121.632877] env[61986]: DEBUG nova.compute.manager [-] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1121.632972] env[61986]: DEBUG nova.network.neutron [-] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1121.907215] env[61986]: DEBUG nova.compute.manager [req-e538a6f0-889e-4725-b5d0-59e9c9c7bd6b req-43f462a7-9bb7-44ea-9e45-59dcda23ebcc service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Received event network-vif-deleted-1146e055-1ec5-4dff-9b99-ec02b64b7cfb {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1121.907215] env[61986]: INFO nova.compute.manager [req-e538a6f0-889e-4725-b5d0-59e9c9c7bd6b req-43f462a7-9bb7-44ea-9e45-59dcda23ebcc service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Neutron deleted interface 1146e055-1ec5-4dff-9b99-ec02b64b7cfb; detaching it from the instance and deleting it from the info cache [ 1121.907215] env[61986]: DEBUG nova.network.neutron [req-e538a6f0-889e-4725-b5d0-59e9c9c7bd6b req-43f462a7-9bb7-44ea-9e45-59dcda23ebcc service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.945973] env[61986]: DEBUG nova.scheduler.client.report [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1122.379253] env[61986]: DEBUG nova.network.neutron [-] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.410892] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9cb1ec16-f41c-4175-a52e-3ed3355c259f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.420814] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331be051-7202-4a92-81e9-7e3602c9ed00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.435802] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.436037] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.454612] env[61986]: DEBUG nova.compute.manager [req-e538a6f0-889e-4725-b5d0-59e9c9c7bd6b req-43f462a7-9bb7-44ea-9e45-59dcda23ebcc service nova] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Detach interface failed, port_id=1146e055-1ec5-4dff-9b99-ec02b64b7cfb, reason: Instance fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1122.825546] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.825844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.882144] env[61986]: INFO nova.compute.manager [-] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Took 1.25 seconds to deallocate network for instance. [ 1122.938106] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1122.957469] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.677s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.328552] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1123.388360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.388643] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.388889] env[61986]: DEBUG nova.objects.instance [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1123.464630] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.848743] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.985969] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6fda7d-2049-4166-aae0-5977598f943d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.993866] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a07fdb-1af0-4194-8cb6-c9a8fe495f54 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.024473] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d7e7ee-86f1-4d05-bc2c-c6dc91935ade {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.032065] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a83636-16f8-404c-b53b-69341d5b8443 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.046199] env[61986]: DEBUG nova.compute.provider_tree [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.493225] env[61986]: INFO nova.compute.manager [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Swapping old allocation on dict_keys(['2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d']) held by migration bdfc7468-370a-4792-b625-c9cb7c123f71 for instance [ 1124.514174] env[61986]: DEBUG nova.scheduler.client.report [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Overwriting current allocation {'allocations': {'2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 140}}, 'project_id': 'c88c612fe00a4ad589db67f3016dace6', 'user_id': '3f3481412ab54a0fb386a6d044a9ae7c', 'consumer_generation': 1} on consumer e825bf09-7f6c-45db-b6cf-9f2dabb98677 {{(pid=61986) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1124.549109] env[61986]: DEBUG nova.scheduler.client.report [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.588226] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.588435] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.588614] env[61986]: DEBUG nova.network.neutron [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1125.054171] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.056619] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.592s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.057957] env[61986]: INFO nova.compute.claims [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1125.072467] env[61986]: INFO nova.scheduler.client.report [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff [ 1125.300492] env[61986]: DEBUG nova.network.neutron [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [{"id": "1a9aae37-5d10-441b-a606-1d2451e00510", "address": "fa:16:3e:55:fb:6d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a9aae37-5d", "ovs_interfaceid": "1a9aae37-5d10-441b-a606-1d2451e00510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.579684] env[61986]: DEBUG oslo_concurrency.lockutils [None req-cca3fd8c-686c-4bc5-9afc-ee69fcfaf5a3 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.058s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.803709] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-e825bf09-7f6c-45db-b6cf-9f2dabb98677" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.804757] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e6a0ff-46ef-4540-898d-c61bb2756db8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.812173] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25498c4-1210-4e8d-b0b3-a717fac0f53a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.149746] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9d92b3-8c7a-484e-917d-9619263acf51 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.157425] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842ecedf-2a09-4bdd-9b0b-de53a740254f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.187291] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0278450-07de-4dc3-906a-fcd0266ee5ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.193993] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feddf805-1779-4917-86ba-34af76f738e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.206382] env[61986]: DEBUG nova.compute.provider_tree [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.350162] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "02950912-c04e-4631-ad06-7f57495f00ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.350401] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "02950912-c04e-4631-ad06-7f57495f00ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.709295] env[61986]: DEBUG nova.scheduler.client.report [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.852383] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1126.895741] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1126.896060] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-227cfd59-d1c8-451b-959f-4c42e2540596 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.903291] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1126.903291] env[61986]: value = "task-1160511" [ 1126.903291] env[61986]: _type = "Task" [ 1126.903291] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.911311] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.214920] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.215495] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.218768] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.370s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.220291] env[61986]: INFO nova.compute.claims [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.374428] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.412820] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160511, 'name': PowerOffVM_Task, 'duration_secs': 0.20779} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.413088] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1127.413719] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.413932] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.414106] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.414293] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.414441] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.414592] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.414794] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.414958] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.415140] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.415309] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.415480] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.420383] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-571f6100-f86b-43b6-b637-a46209bd1ce8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.434621] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1127.434621] env[61986]: value = "task-1160512" [ 1127.434621] env[61986]: _type = "Task" [ 1127.434621] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.445124] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160512, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.720648] env[61986]: DEBUG nova.compute.utils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.722193] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.722354] env[61986]: DEBUG nova.network.neutron [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1127.762927] env[61986]: DEBUG nova.policy [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4b4b1412f644b6097f4de846388dec1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '691d4935610e45e6a0bae5ca5f1197f3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1127.944619] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160512, 'name': ReconfigVM_Task, 'duration_secs': 0.153796} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.945460] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb757ed-8b48-44a7-bebc-259063f8d15e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.968615] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.968864] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.969489] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.969489] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.969489] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.969609] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.969779] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.969943] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.970126] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.970292] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.970464] env[61986]: DEBUG nova.virt.hardware [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.971294] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a0ff484-2ea7-4c88-9d44-26caca7c6b44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.977212] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1127.977212] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a9bc42-9e6d-cbe7-b35b-6a0244fe09b3" [ 1127.977212] env[61986]: _type = "Task" [ 1127.977212] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.985280] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a9bc42-9e6d-cbe7-b35b-6a0244fe09b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.038423] env[61986]: DEBUG nova.network.neutron [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Successfully created port: 0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1128.225576] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1128.324819] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3596fc-a23f-439a-a169-966a7578c95a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.332396] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9edfb7b-0db4-4321-8024-16e548dc29fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.361543] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dceb6dc5-35bf-4cc6-86fe-9697dea1eb2d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.368553] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645f8fe4-b562-4cb6-a2bf-4699929d3210 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.382553] env[61986]: DEBUG nova.compute.provider_tree [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.487453] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a9bc42-9e6d-cbe7-b35b-6a0244fe09b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008211} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.492680] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1128.492985] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-012a83a3-3904-4f0b-bd92-dbd41d4f9a43 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.510238] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1128.510238] env[61986]: value = "task-1160513" [ 1128.510238] env[61986]: _type = "Task" [ 1128.510238] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.517587] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160513, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.885880] env[61986]: DEBUG nova.scheduler.client.report [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1129.020480] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160513, 'name': ReconfigVM_Task, 'duration_secs': 0.244896} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.020696] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1129.021505] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcc5a69-79b7-4e17-b9b9-6eaf5bd3bc3f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.046165] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1129.046435] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-705fe5b5-dfa9-4626-9aa0-9d49e037d30c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.064539] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1129.064539] env[61986]: value = "task-1160514" [ 1129.064539] env[61986]: _type = "Task" [ 1129.064539] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.071762] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160514, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.237995] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1129.264853] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.265200] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.265426] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.265700] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.265900] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.266105] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.266363] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.266567] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.266780] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.266990] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.267230] env[61986]: DEBUG nova.virt.hardware [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.268199] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240df607-1f29-495b-9418-0dc5028114d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.276933] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bdac2f-db92-4427-b819-670406a93af3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.391267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.391849] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1129.394590] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.020s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.395978] env[61986]: INFO nova.compute.claims [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1129.427269] env[61986]: DEBUG nova.compute.manager [req-bbd00798-8c8b-4fb9-9331-b848b5344b54 req-73eeaa6a-22fc-42f6-a850-6fef970aaf17 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-vif-plugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1129.427604] env[61986]: DEBUG oslo_concurrency.lockutils [req-bbd00798-8c8b-4fb9-9331-b848b5344b54 req-73eeaa6a-22fc-42f6-a850-6fef970aaf17 service nova] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.427731] env[61986]: DEBUG oslo_concurrency.lockutils [req-bbd00798-8c8b-4fb9-9331-b848b5344b54 req-73eeaa6a-22fc-42f6-a850-6fef970aaf17 service nova] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.427912] env[61986]: DEBUG oslo_concurrency.lockutils [req-bbd00798-8c8b-4fb9-9331-b848b5344b54 req-73eeaa6a-22fc-42f6-a850-6fef970aaf17 service nova] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.428174] env[61986]: DEBUG nova.compute.manager [req-bbd00798-8c8b-4fb9-9331-b848b5344b54 req-73eeaa6a-22fc-42f6-a850-6fef970aaf17 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] No waiting events found dispatching network-vif-plugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.428377] env[61986]: WARNING nova.compute.manager [req-bbd00798-8c8b-4fb9-9331-b848b5344b54 req-73eeaa6a-22fc-42f6-a850-6fef970aaf17 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received unexpected event network-vif-plugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f for instance with vm_state building and task_state spawning. [ 1129.518260] env[61986]: DEBUG nova.network.neutron [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Successfully updated port: 0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.574274] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160514, 'name': ReconfigVM_Task, 'duration_secs': 0.291634} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.574553] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to attach disk [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677/e825bf09-7f6c-45db-b6cf-9f2dabb98677.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.575389] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6681bb0-ab00-4257-9a01-327ec6ffb5a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.596611] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec853867-9a3a-4a4f-902b-0e639765bdc5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.616225] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e7e7f1-e612-4c31-8531-03820473e049 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.635797] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4beecf7-81a6-42ff-8cef-4058e37e5af4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.642396] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1129.642639] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3a8de2a-98fd-44d7-ab2c-18916e83bd62 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.648241] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1129.648241] env[61986]: value = "task-1160515" [ 1129.648241] env[61986]: _type = "Task" [ 1129.648241] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.655266] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.900119] env[61986]: DEBUG nova.compute.utils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1129.903637] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1129.903637] env[61986]: DEBUG nova.network.neutron [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1129.940472] env[61986]: DEBUG nova.policy [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c40c9363dc7741a3bd6a040f20284837', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e0f7d6b6be498eaec797aa64aaea7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1130.021934] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.021934] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.022106] env[61986]: DEBUG nova.network.neutron [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1130.160688] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160515, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.177775] env[61986]: DEBUG nova.network.neutron [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Successfully created port: 9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1130.406678] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1130.508072] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac169b3-568e-4d40-b1ca-df2c4729b04e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.515613] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe4e206-ff4c-4a20-895a-4cb520194bcf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.547572] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d9c1ca-b19e-45e1-93b9-0560b4e67e3d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.554999] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f291f3a-dc6a-4f16-bc29-e3d07ae2e44e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.568379] env[61986]: DEBUG nova.compute.provider_tree [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.585904] env[61986]: DEBUG nova.network.neutron [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.660270] env[61986]: DEBUG oslo_vmware.api [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160515, 'name': PowerOnVM_Task, 'duration_secs': 1.005544} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.660535] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1130.726419] env[61986]: DEBUG nova.network.neutron [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.071469] env[61986]: DEBUG nova.scheduler.client.report [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.228731] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.229154] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Instance network_info: |[{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1131.229629] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:0e:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1768af3d-3317-4ef5-b484-0c2707d63de7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b5285ba-76dd-4e56-aa09-5ac60d84235f', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.237151] env[61986]: DEBUG oslo.service.loopingcall [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.237572] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1131.237797] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-796a3c54-ec94-40b7-89e9-bff5eaf1a739 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.257739] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.257739] env[61986]: value = "task-1160516" [ 1131.257739] env[61986]: _type = "Task" [ 1131.257739] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.265160] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160516, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.419973] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1131.448583] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1131.448857] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1131.449060] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1131.449217] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1131.449368] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1131.449568] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1131.449822] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1131.449956] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1131.450193] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1131.450379] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1131.450633] env[61986]: DEBUG nova.virt.hardware [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1131.451682] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ca0013-1cf9-47fe-bc30-9ea645fd681d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.456732] env[61986]: DEBUG nova.compute.manager [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.456946] env[61986]: DEBUG nova.compute.manager [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing instance network info cache due to event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1131.457284] env[61986]: DEBUG oslo_concurrency.lockutils [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.457454] env[61986]: DEBUG oslo_concurrency.lockutils [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.457621] env[61986]: DEBUG nova.network.neutron [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1131.464863] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f2ab9d-4c8d-40a8-a92d-ead67c7134ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.576898] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.182s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.577510] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1131.636115] env[61986]: DEBUG nova.network.neutron [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Successfully updated port: 9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1131.707883] env[61986]: INFO nova.compute.manager [None req-f1b77782-7d4d-4f83-90ea-086959428ecc tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance to original state: 'active' [ 1131.767225] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160516, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.081862] env[61986]: DEBUG nova.compute.utils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1132.083260] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1132.084260] env[61986]: DEBUG nova.network.neutron [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1132.140785] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-43cb9710-3f7d-400f-b046-16c0c77e26b7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.140785] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-43cb9710-3f7d-400f-b046-16c0c77e26b7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.140785] env[61986]: DEBUG nova.network.neutron [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1132.146189] env[61986]: DEBUG nova.policy [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1132.199395] env[61986]: DEBUG nova.network.neutron [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updated VIF entry in instance network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1132.199395] env[61986]: DEBUG nova.network.neutron [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.271872] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160516, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.412194] env[61986]: DEBUG nova.network.neutron [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Successfully created port: 2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1132.587521] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1132.671025] env[61986]: DEBUG nova.network.neutron [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1132.702447] env[61986]: DEBUG oslo_concurrency.lockutils [req-3e21113b-1045-41b3-81b8-c0e17d71aa67 req-3b625cbd-8a3d-43de-a6c6-a6e2b84df01c service nova] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.769914] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160516, 'name': CreateVM_Task, 'duration_secs': 1.294421} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.771080] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1132.774011] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.774011] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.774323] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1132.774586] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d9600fe-12b9-4f70-9d7a-7baa5fbed596 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.779537] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1132.779537] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a02e74-4fa8-a61f-3f07-3b4f6a918405" [ 1132.779537] env[61986]: _type = "Task" [ 1132.779537] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.787201] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a02e74-4fa8-a61f-3f07-3b4f6a918405, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.820411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.820677] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.820903] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.821246] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.821363] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.823500] env[61986]: INFO nova.compute.manager [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Terminating instance [ 1132.825497] env[61986]: DEBUG nova.compute.manager [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1132.825716] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1132.825957] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40b65525-b734-41ee-a78f-46ea03933837 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.832236] env[61986]: DEBUG nova.network.neutron [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Updating instance_info_cache with network_info: [{"id": "9741a942-9250-4623-a520-3bf6d16b4f69", "address": "fa:16:3e:0b:3c:37", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9741a942-92", "ovs_interfaceid": "9741a942-9250-4623-a520-3bf6d16b4f69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.834806] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1132.834806] env[61986]: value = "task-1160517" [ 1132.834806] env[61986]: _type = "Task" [ 1132.834806] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.843908] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160517, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.291190] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a02e74-4fa8-a61f-3f07-3b4f6a918405, 'name': SearchDatastore_Task, 'duration_secs': 0.008802} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.291524] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.291765] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.292011] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.292240] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.292426] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.292695] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d7fb44c-cb10-4045-81d7-5be985cc019f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.300850] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.301073] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1133.301747] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35a1b7f7-8151-4558-a63f-43b272ded861 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.306942] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1133.306942] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5240ab88-713b-78b6-cce8-4e96d23e63b2" [ 1133.306942] env[61986]: _type = "Task" [ 1133.306942] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.314415] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5240ab88-713b-78b6-cce8-4e96d23e63b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.333984] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-43cb9710-3f7d-400f-b046-16c0c77e26b7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.334297] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Instance network_info: |[{"id": "9741a942-9250-4623-a520-3bf6d16b4f69", "address": "fa:16:3e:0b:3c:37", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9741a942-92", "ovs_interfaceid": "9741a942-9250-4623-a520-3bf6d16b4f69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1133.334671] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:3c:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9741a942-9250-4623-a520-3bf6d16b4f69', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1133.342159] env[61986]: DEBUG oslo.service.loopingcall [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.342429] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1133.345329] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be9ca2f1-1b83-41f2-a6a0-e2bd60bbacab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.364441] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160517, 'name': PowerOffVM_Task, 'duration_secs': 0.19675} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.365578] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.365785] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1133.365976] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252470', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'name': 'volume-1568453a-0d54-4730-b64d-cb77ca01e291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'e825bf09-7f6c-45db-b6cf-9f2dabb98677', 'attached_at': '2024-10-07T06:48:01.000000', 'detached_at': '', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'serial': '1568453a-0d54-4730-b64d-cb77ca01e291'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1133.366221] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1133.366221] env[61986]: value = "task-1160518" [ 1133.366221] env[61986]: _type = "Task" [ 1133.366221] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.366863] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc1c627-c861-4976-95dd-394c7fd785e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.393631] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb06260-88e6-4271-9efa-79d150328cf9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.395975] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160518, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.400127] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38e3c59-0b1f-4c00-bd83-88b5d0d7ba97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.421034] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e6393f-bb07-4e9e-bd35-9b0abfaa8a1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.436487] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] The volume has not been displaced from its original location: [datastore2] volume-1568453a-0d54-4730-b64d-cb77ca01e291/volume-1568453a-0d54-4730-b64d-cb77ca01e291.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1133.441806] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.442195] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b9bb223-1581-4444-bbb3-27c292260a12 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.460361] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1133.460361] env[61986]: value = "task-1160519" [ 1133.460361] env[61986]: _type = "Task" [ 1133.460361] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.468323] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160519, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.486031] env[61986]: DEBUG nova.compute.manager [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Received event network-vif-plugged-9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.486350] env[61986]: DEBUG oslo_concurrency.lockutils [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] Acquiring lock "43cb9710-3f7d-400f-b046-16c0c77e26b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.486529] env[61986]: DEBUG oslo_concurrency.lockutils [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.486715] env[61986]: DEBUG oslo_concurrency.lockutils [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.486883] env[61986]: DEBUG nova.compute.manager [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] No waiting events found dispatching network-vif-plugged-9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1133.487089] env[61986]: WARNING nova.compute.manager [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Received unexpected event network-vif-plugged-9741a942-9250-4623-a520-3bf6d16b4f69 for instance with vm_state building and task_state spawning. [ 1133.487342] env[61986]: DEBUG nova.compute.manager [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Received event network-changed-9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.487415] env[61986]: DEBUG nova.compute.manager [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Refreshing instance network info cache due to event network-changed-9741a942-9250-4623-a520-3bf6d16b4f69. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1133.487597] env[61986]: DEBUG oslo_concurrency.lockutils [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] Acquiring lock "refresh_cache-43cb9710-3f7d-400f-b046-16c0c77e26b7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.487735] env[61986]: DEBUG oslo_concurrency.lockutils [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] Acquired lock "refresh_cache-43cb9710-3f7d-400f-b046-16c0c77e26b7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.487893] env[61986]: DEBUG nova.network.neutron [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Refreshing network info cache for port 9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1133.596460] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1133.622792] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1133.623118] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1133.623341] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.623569] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1133.623756] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.623950] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1133.624228] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1133.624438] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1133.624649] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1133.624854] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1133.625084] env[61986]: DEBUG nova.virt.hardware [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.626058] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aeb310b-f7b3-4610-b2f1-30d00ee0fe3c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.634734] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be14d93-b6c2-4199-9174-0ff5de337df5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.819626] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5240ab88-713b-78b6-cce8-4e96d23e63b2, 'name': SearchDatastore_Task, 'duration_secs': 0.008891} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.820513] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0aeb6f5-d921-4a1f-8755-e90717397899 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.825837] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1133.825837] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528ebe45-9874-f0a8-c60a-07579f3ef9a5" [ 1133.825837] env[61986]: _type = "Task" [ 1133.825837] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.835313] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528ebe45-9874-f0a8-c60a-07579f3ef9a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.880616] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160518, 'name': CreateVM_Task, 'duration_secs': 0.315288} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.881630] env[61986]: DEBUG nova.network.neutron [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Successfully updated port: 2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1133.882744] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1133.884779] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.884779] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.884779] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1133.884996] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f43297-d544-4da7-ae53-404b2c86a4e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.889560] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1133.889560] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52967efc-b729-4d1d-8b4f-83089901ab84" [ 1133.889560] env[61986]: _type = "Task" [ 1133.889560] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.898011] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52967efc-b729-4d1d-8b4f-83089901ab84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.970072] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160519, 'name': ReconfigVM_Task, 'duration_secs': 0.234057} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.970390] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.974974] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41b2c348-50f5-4714-9712-a97309b80e5c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.989460] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1133.989460] env[61986]: value = "task-1160520" [ 1133.989460] env[61986]: _type = "Task" [ 1133.989460] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.998796] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.337926] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528ebe45-9874-f0a8-c60a-07579f3ef9a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008711} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.337926] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.337926] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1134.337926] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71e9bc8d-8493-451f-959e-d8c785c5b9ca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.345030] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1134.345030] env[61986]: value = "task-1160521" [ 1134.345030] env[61986]: _type = "Task" [ 1134.345030] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.352454] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160521, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.359545] env[61986]: DEBUG nova.network.neutron [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Updated VIF entry in instance network info cache for port 9741a942-9250-4623-a520-3bf6d16b4f69. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1134.359881] env[61986]: DEBUG nova.network.neutron [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Updating instance_info_cache with network_info: [{"id": "9741a942-9250-4623-a520-3bf6d16b4f69", "address": "fa:16:3e:0b:3c:37", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9741a942-92", "ovs_interfaceid": "9741a942-9250-4623-a520-3bf6d16b4f69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.386077] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-02950912-c04e-4631-ad06-7f57495f00ea" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.386218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-02950912-c04e-4631-ad06-7f57495f00ea" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.386364] env[61986]: DEBUG nova.network.neutron [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1134.400144] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52967efc-b729-4d1d-8b4f-83089901ab84, 'name': SearchDatastore_Task, 'duration_secs': 0.02827} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.400426] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.400646] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1134.400871] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.401053] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.401249] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1134.401522] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17fef736-1caf-4d7f-bac0-81d953f01b87 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.408698] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1134.408880] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1134.409573] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c769a3c2-2ef2-4ffc-a5b9-0c53bbc3f7fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.415316] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1134.415316] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b571bb-cf3c-1889-9997-5e66015e290a" [ 1134.415316] env[61986]: _type = "Task" [ 1134.415316] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.422375] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b571bb-cf3c-1889-9997-5e66015e290a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.499336] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160520, 'name': ReconfigVM_Task, 'duration_secs': 0.132695} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.499624] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252470', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'name': 'volume-1568453a-0d54-4730-b64d-cb77ca01e291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'e825bf09-7f6c-45db-b6cf-9f2dabb98677', 'attached_at': '2024-10-07T06:48:01.000000', 'detached_at': '', 'volume_id': '1568453a-0d54-4730-b64d-cb77ca01e291', 'serial': '1568453a-0d54-4730-b64d-cb77ca01e291'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1134.499919] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1134.500760] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f47f3a2-1b2c-48b0-b331-bfb76d40f3ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.507763] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1134.508048] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2906bb77-7d6e-4946-84f5-bceba674ab17 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.575986] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1134.576261] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1134.576469] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleting the datastore file [datastore1] e825bf09-7f6c-45db-b6cf-9f2dabb98677 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1134.576741] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eec5130e-e13a-4c09-83d0-af2d02af46e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.584938] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1134.584938] env[61986]: value = "task-1160523" [ 1134.584938] env[61986]: _type = "Task" [ 1134.584938] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.593191] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160523, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.854123] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160521, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484306} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.854391] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1134.854609] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.854868] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8bdd9b26-a9a3-4749-b8e3-4931b80bc39e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.861124] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1134.861124] env[61986]: value = "task-1160524" [ 1134.861124] env[61986]: _type = "Task" [ 1134.861124] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.865150] env[61986]: DEBUG oslo_concurrency.lockutils [req-970aed66-0ef7-4ebe-b711-bef5b897509d req-56c3fd67-e73f-4ebd-bfb0-b7d4261bcb02 service nova] Releasing lock "refresh_cache-43cb9710-3f7d-400f-b046-16c0c77e26b7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.869874] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.928182] env[61986]: DEBUG nova.network.neutron [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1134.936018] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b571bb-cf3c-1889-9997-5e66015e290a, 'name': SearchDatastore_Task, 'duration_secs': 0.008265} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.936960] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e319642-164c-44bb-92ea-2d1e99cbf8d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.942686] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1134.942686] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a17226-7c54-e7e6-f8da-e021a55e66d2" [ 1134.942686] env[61986]: _type = "Task" [ 1134.942686] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.952404] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a17226-7c54-e7e6-f8da-e021a55e66d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.059813] env[61986]: DEBUG nova.network.neutron [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Updating instance_info_cache with network_info: [{"id": "2219b83b-06dd-449f-9274-be9d72376018", "address": "fa:16:3e:bc:fa:9a", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2219b83b-06", "ovs_interfaceid": "2219b83b-06dd-449f-9274-be9d72376018", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.095354] env[61986]: DEBUG oslo_vmware.api [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160523, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339185} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.095655] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1135.095890] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1135.096106] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1135.096288] env[61986]: INFO nova.compute.manager [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Took 2.27 seconds to destroy the instance on the hypervisor. [ 1135.096751] env[61986]: DEBUG oslo.service.loopingcall [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1135.096751] env[61986]: DEBUG nova.compute.manager [-] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1135.096827] env[61986]: DEBUG nova.network.neutron [-] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1135.372013] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065371} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.372433] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1135.373219] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b76de2-baa6-4879-bd2f-d0bdbd2cf7e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.395415] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.395710] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-165adb38-2ca6-4a25-937b-5e0f9a27c156 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.417981] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1135.417981] env[61986]: value = "task-1160525" [ 1135.417981] env[61986]: _type = "Task" [ 1135.417981] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.425618] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.453821] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52a17226-7c54-e7e6-f8da-e021a55e66d2, 'name': SearchDatastore_Task, 'duration_secs': 0.015434} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.454414] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.454751] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 43cb9710-3f7d-400f-b046-16c0c77e26b7/43cb9710-3f7d-400f-b046-16c0c77e26b7.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1135.455110] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-106c9de5-ee88-4d83-ab1d-5b19ba948f2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.463302] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1135.463302] env[61986]: value = "task-1160526" [ 1135.463302] env[61986]: _type = "Task" [ 1135.463302] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.472667] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.518019] env[61986]: DEBUG nova.compute.manager [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Received event network-vif-plugged-2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.518019] env[61986]: DEBUG oslo_concurrency.lockutils [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] Acquiring lock "02950912-c04e-4631-ad06-7f57495f00ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.518019] env[61986]: DEBUG oslo_concurrency.lockutils [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] Lock "02950912-c04e-4631-ad06-7f57495f00ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.518019] env[61986]: DEBUG oslo_concurrency.lockutils [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] Lock "02950912-c04e-4631-ad06-7f57495f00ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.518019] env[61986]: DEBUG nova.compute.manager [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] No waiting events found dispatching network-vif-plugged-2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1135.518019] env[61986]: WARNING nova.compute.manager [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Received unexpected event network-vif-plugged-2219b83b-06dd-449f-9274-be9d72376018 for instance with vm_state building and task_state spawning. [ 1135.518019] env[61986]: DEBUG nova.compute.manager [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Received event network-changed-2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.518019] env[61986]: DEBUG nova.compute.manager [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Refreshing instance network info cache due to event network-changed-2219b83b-06dd-449f-9274-be9d72376018. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1135.518019] env[61986]: DEBUG oslo_concurrency.lockutils [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] Acquiring lock "refresh_cache-02950912-c04e-4631-ad06-7f57495f00ea" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.563073] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-02950912-c04e-4631-ad06-7f57495f00ea" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.563073] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Instance network_info: |[{"id": "2219b83b-06dd-449f-9274-be9d72376018", "address": "fa:16:3e:bc:fa:9a", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2219b83b-06", "ovs_interfaceid": "2219b83b-06dd-449f-9274-be9d72376018", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1135.563073] env[61986]: DEBUG oslo_concurrency.lockutils [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] Acquired lock "refresh_cache-02950912-c04e-4631-ad06-7f57495f00ea" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.563361] env[61986]: DEBUG nova.network.neutron [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Refreshing network info cache for port 2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1135.564421] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:fa:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2219b83b-06dd-449f-9274-be9d72376018', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1135.572553] env[61986]: DEBUG oslo.service.loopingcall [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1135.573503] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1135.573740] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8769526c-64c0-49f5-882f-152774dfc7ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.594673] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1135.594673] env[61986]: value = "task-1160527" [ 1135.594673] env[61986]: _type = "Task" [ 1135.594673] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.603444] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160527, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.928665] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160525, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.975054] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160526, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.033732] env[61986]: DEBUG nova.network.neutron [-] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.104038] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160527, 'name': CreateVM_Task, 'duration_secs': 0.382256} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.106200] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1136.106856] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.107037] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.107360] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1136.107893] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96febffb-b7d6-4159-bba4-283f168464cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.113733] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1136.113733] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b330ee-86ac-4522-baf8-5b9335c34aee" [ 1136.113733] env[61986]: _type = "Task" [ 1136.113733] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.121617] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b330ee-86ac-4522-baf8-5b9335c34aee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.268843] env[61986]: DEBUG nova.network.neutron [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Updated VIF entry in instance network info cache for port 2219b83b-06dd-449f-9274-be9d72376018. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1136.269300] env[61986]: DEBUG nova.network.neutron [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Updating instance_info_cache with network_info: [{"id": "2219b83b-06dd-449f-9274-be9d72376018", "address": "fa:16:3e:bc:fa:9a", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2219b83b-06", "ovs_interfaceid": "2219b83b-06dd-449f-9274-be9d72376018", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.428553] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160525, 'name': ReconfigVM_Task, 'duration_secs': 0.531754} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.428810] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.429582] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd7885de-9f7b-4f68-ba4f-f6f6da9d1e5d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.435796] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1136.435796] env[61986]: value = "task-1160528" [ 1136.435796] env[61986]: _type = "Task" [ 1136.435796] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.443052] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160528, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.472319] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587461} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.472559] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 43cb9710-3f7d-400f-b046-16c0c77e26b7/43cb9710-3f7d-400f-b046-16c0c77e26b7.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1136.472773] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1136.473018] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4aa580d3-50f9-4db0-8789-66080d0f29be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.479805] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1136.479805] env[61986]: value = "task-1160529" [ 1136.479805] env[61986]: _type = "Task" [ 1136.479805] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.487591] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.536776] env[61986]: INFO nova.compute.manager [-] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Took 1.44 seconds to deallocate network for instance. [ 1136.624159] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b330ee-86ac-4522-baf8-5b9335c34aee, 'name': SearchDatastore_Task, 'duration_secs': 0.008208} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.624463] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.624710] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1136.624948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.625114] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.625297] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1136.625557] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-116458b8-0a4e-4c3c-8213-c8ca13a801f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.632600] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1136.632777] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1136.633543] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cb19870-d34b-4f96-b78a-628436c69560 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.638218] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1136.638218] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b2aeaa-6c0d-3a46-b7ee-953dbce240a5" [ 1136.638218] env[61986]: _type = "Task" [ 1136.638218] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.644921] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b2aeaa-6c0d-3a46-b7ee-953dbce240a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.771939] env[61986]: DEBUG oslo_concurrency.lockutils [req-329ef4fe-7574-4be0-a8ad-7f3cdb43d27e req-126a9547-e92a-44bc-b11a-1d2249688ed8 service nova] Releasing lock "refresh_cache-02950912-c04e-4631-ad06-7f57495f00ea" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.945866] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160528, 'name': Rename_Task, 'duration_secs': 0.153208} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.946181] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1136.946413] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-79152da4-f326-484b-9b1f-fe090cd5b4db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.952200] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1136.952200] env[61986]: value = "task-1160530" [ 1136.952200] env[61986]: _type = "Task" [ 1136.952200] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.959342] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160530, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.989543] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063651} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.989787] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1136.990595] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613f3a52-de91-407f-b412-55a76884cbc7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.013095] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 43cb9710-3f7d-400f-b046-16c0c77e26b7/43cb9710-3f7d-400f-b046-16c0c77e26b7.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.013415] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9966e1cc-5408-40ca-a222-d2d5085ec0ef {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.032447] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1137.032447] env[61986]: value = "task-1160531" [ 1137.032447] env[61986]: _type = "Task" [ 1137.032447] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.041472] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160531, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.082509] env[61986]: INFO nova.compute.manager [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Took 0.55 seconds to detach 1 volumes for instance. [ 1137.150050] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b2aeaa-6c0d-3a46-b7ee-953dbce240a5, 'name': SearchDatastore_Task, 'duration_secs': 0.009694} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.150948] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c09c8c00-f043-42e6-83b7-69c3877aef54 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.156661] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1137.156661] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523c959d-cd0b-0921-874e-0a6eb4fca70b" [ 1137.156661] env[61986]: _type = "Task" [ 1137.156661] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.164260] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523c959d-cd0b-0921-874e-0a6eb4fca70b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.461654] env[61986]: DEBUG oslo_vmware.api [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160530, 'name': PowerOnVM_Task, 'duration_secs': 0.483388} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.461932] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1137.462159] env[61986]: INFO nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Took 8.22 seconds to spawn the instance on the hypervisor. [ 1137.462347] env[61986]: DEBUG nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1137.463120] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f39418-1310-4b18-ada2-b9c8bd8da9e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.541871] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160531, 'name': ReconfigVM_Task, 'duration_secs': 0.3404} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.543135] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 43cb9710-3f7d-400f-b046-16c0c77e26b7/43cb9710-3f7d-400f-b046-16c0c77e26b7.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.544595] env[61986]: DEBUG nova.compute.manager [req-f7a9d049-e8e6-4eec-8440-f22e5d75ee27 req-81211c75-4ee6-4320-964c-fc128a36ff8b service nova] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Received event network-vif-deleted-1a9aae37-5d10-441b-a606-1d2451e00510 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1137.544900] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecabe1ae-013e-4909-8a6e-3f4b3a99e398 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.551201] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1137.551201] env[61986]: value = "task-1160532" [ 1137.551201] env[61986]: _type = "Task" [ 1137.551201] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.562618] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160532, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.588494] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.588732] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.588916] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.615353] env[61986]: INFO nova.scheduler.client.report [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted allocations for instance e825bf09-7f6c-45db-b6cf-9f2dabb98677 [ 1137.669334] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]523c959d-cd0b-0921-874e-0a6eb4fca70b, 'name': SearchDatastore_Task, 'duration_secs': 0.029579} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.669334] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.669556] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 02950912-c04e-4631-ad06-7f57495f00ea/02950912-c04e-4631-ad06-7f57495f00ea.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1137.669680] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa265f21-30fa-402a-baeb-7432beda6360 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.677244] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1137.677244] env[61986]: value = "task-1160533" [ 1137.677244] env[61986]: _type = "Task" [ 1137.677244] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.684926] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.982155] env[61986]: INFO nova.compute.manager [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Took 14.54 seconds to build instance. [ 1138.061608] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160532, 'name': Rename_Task, 'duration_secs': 0.195471} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.061971] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1138.062604] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-064437fa-0acd-4364-a6e6-874a3616bea6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.070166] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1138.070166] env[61986]: value = "task-1160534" [ 1138.070166] env[61986]: _type = "Task" [ 1138.070166] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.078424] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.123950] env[61986]: DEBUG oslo_concurrency.lockutils [None req-22938dcf-a4bd-4183-8575-349135b9d782 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "e825bf09-7f6c-45db-b6cf-9f2dabb98677" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.303s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.188530] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476161} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.188820] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 02950912-c04e-4631-ad06-7f57495f00ea/02950912-c04e-4631-ad06-7f57495f00ea.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1138.189055] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1138.189313] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-998cee70-ca65-4cee-8cd0-0f62c95462e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.195741] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1138.195741] env[61986]: value = "task-1160535" [ 1138.195741] env[61986]: _type = "Task" [ 1138.195741] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.203603] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160535, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.484904] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4243c5a2-c421-4960-b9d7-a8f4bb49f472 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.049s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.580537] env[61986]: DEBUG oslo_vmware.api [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160534, 'name': PowerOnVM_Task, 'duration_secs': 0.455462} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.580537] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1138.580730] env[61986]: INFO nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Took 7.16 seconds to spawn the instance on the hypervisor. [ 1138.580907] env[61986]: DEBUG nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.581685] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f44ef65-c7e0-459d-a00c-a6bee7c2d726 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.705271] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160535, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061168} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.705561] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1138.706392] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e302586-067f-48c9-b74a-99a34c4ed9c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.729220] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 02950912-c04e-4631-ad06-7f57495f00ea/02950912-c04e-4631-ad06-7f57495f00ea.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1138.730024] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a20dfbd4-662f-4881-a11e-35f222bf628d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.750508] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1138.750508] env[61986]: value = "task-1160536" [ 1138.750508] env[61986]: _type = "Task" [ 1138.750508] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.758750] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160536, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.955162] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.955396] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.053204] env[61986]: DEBUG nova.compute.manager [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1139.053473] env[61986]: DEBUG nova.compute.manager [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing instance network info cache due to event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1139.053651] env[61986]: DEBUG oslo_concurrency.lockutils [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.053799] env[61986]: DEBUG oslo_concurrency.lockutils [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.053962] env[61986]: DEBUG nova.network.neutron [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1139.097025] env[61986]: INFO nova.compute.manager [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Took 15.26 seconds to build instance. [ 1139.260149] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160536, 'name': ReconfigVM_Task, 'duration_secs': 0.252632} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.260464] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 02950912-c04e-4631-ad06-7f57495f00ea/02950912-c04e-4631-ad06-7f57495f00ea.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.261135] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8113416-b668-45f9-917d-2bbd4864722e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.268343] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1139.268343] env[61986]: value = "task-1160537" [ 1139.268343] env[61986]: _type = "Task" [ 1139.268343] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.277427] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160537, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.457454] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1139.599763] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3e679d6c-fd9a-4210-aff5-d45723895f57 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.774s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.777733] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160537, 'name': Rename_Task, 'duration_secs': 0.174134} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.778080] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1139.778387] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f5f657a-4938-46ba-a0c3-67d5869b3b79 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.780503] env[61986]: DEBUG nova.network.neutron [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updated VIF entry in instance network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1139.780832] env[61986]: DEBUG nova.network.neutron [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.787064] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1139.787064] env[61986]: value = "task-1160538" [ 1139.787064] env[61986]: _type = "Task" [ 1139.787064] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.793809] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.976014] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.976329] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.977925] env[61986]: INFO nova.compute.claims [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.146842] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "52353614-d3bb-46da-a8fc-ab920442c64d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.146842] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.283901] env[61986]: DEBUG oslo_concurrency.lockutils [req-9532f408-22f0-4c33-bf36-ca2125993acb req-94db3447-f59b-40d0-aca1-c68ae0ebb974 service nova] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.296702] env[61986]: DEBUG oslo_vmware.api [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160538, 'name': PowerOnVM_Task, 'duration_secs': 0.449566} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.297088] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1140.297315] env[61986]: INFO nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Took 6.70 seconds to spawn the instance on the hypervisor. [ 1140.297532] env[61986]: DEBUG nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1140.298318] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf486cd-bb5a-4e46-a0e4-cc1dabf19363 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.327481] env[61986]: DEBUG oslo_concurrency.lockutils [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.327774] env[61986]: DEBUG oslo_concurrency.lockutils [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.327966] env[61986]: DEBUG nova.compute.manager [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1140.328840] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf9c6e7-9c93-43ea-93dd-a29b4a0d73fe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.335806] env[61986]: DEBUG nova.compute.manager [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61986) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1140.336402] env[61986]: DEBUG nova.objects.instance [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'flavor' on Instance uuid 43cb9710-3f7d-400f-b046-16c0c77e26b7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.649549] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1140.815544] env[61986]: INFO nova.compute.manager [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Took 13.46 seconds to build instance. [ 1140.841315] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1140.841576] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6eebfa7c-ddeb-4fab-b852-e0eccdc2d24e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.848727] env[61986]: DEBUG oslo_vmware.api [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1140.848727] env[61986]: value = "task-1160539" [ 1140.848727] env[61986]: _type = "Task" [ 1140.848727] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.856234] env[61986]: DEBUG oslo_vmware.api [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.084534] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eea5d2a-d5b4-421b-9a04-6537dbab6a73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.092539] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfbf1b6-ed1a-4e3e-9195-aaea5147ecbe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.122271] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5e2ee6-df80-429d-adb3-b3ac5772e6c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.129548] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899a3b0a-f1a0-497b-96c5-c5939366bad8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.142090] env[61986]: DEBUG nova.compute.provider_tree [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.167507] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.317259] env[61986]: DEBUG oslo_concurrency.lockutils [None req-71f786eb-df27-4296-b0d5-bb8b4e04c229 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "02950912-c04e-4631-ad06-7f57495f00ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.967s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.358783] env[61986]: DEBUG oslo_vmware.api [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160539, 'name': PowerOffVM_Task, 'duration_secs': 0.395428} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.359112] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1141.359286] env[61986]: DEBUG nova.compute.manager [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1141.360141] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e42727-79d0-457c-833f-d563bd447d37 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.646035] env[61986]: DEBUG nova.scheduler.client.report [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.872878] env[61986]: DEBUG oslo_concurrency.lockutils [None req-db21b2e4-232f-4dd1-a7dd-06f9877acaa3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.046192] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "0250e945-6290-422c-840b-f7afac85331f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.046465] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "0250e945-6290-422c-840b-f7afac85331f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.150387] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.174s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.150858] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1142.153589] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.986s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.154943] env[61986]: INFO nova.compute.claims [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.548956] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1142.659315] env[61986]: DEBUG nova.compute.utils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1142.663075] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1142.663409] env[61986]: DEBUG nova.network.neutron [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1142.709882] env[61986]: DEBUG nova.policy [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f3481412ab54a0fb386a6d044a9ae7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c88c612fe00a4ad589db67f3016dace6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1142.729428] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.729717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.730089] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "43cb9710-3f7d-400f-b046-16c0c77e26b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.730330] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.730528] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.733245] env[61986]: INFO nova.compute.manager [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Terminating instance [ 1142.735220] env[61986]: DEBUG nova.compute.manager [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1142.735310] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1142.736422] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517391a5-4bd6-499f-9a92-176cffbfa158 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.745031] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1142.745031] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73cd5d51-50ce-4f85-90c7-7d130c69b1d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.856908] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1142.857267] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1142.857545] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleting the datastore file [datastore2] 43cb9710-3f7d-400f-b046-16c0c77e26b7 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1142.857871] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0d2f641-300f-4cf6-a6cf-e5e3c604d63c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.865832] env[61986]: DEBUG oslo_vmware.api [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1142.865832] env[61986]: value = "task-1160541" [ 1142.865832] env[61986]: _type = "Task" [ 1142.865832] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.874748] env[61986]: DEBUG oslo_vmware.api [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160541, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.996191] env[61986]: DEBUG nova.network.neutron [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Successfully created port: 37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1143.071838] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.163878] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1143.281981] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98442588-5463-4567-8e62-19d363883127 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.290183] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd3f564-f754-4692-9409-1502ebfe53a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.320981] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e060c77d-004b-4f5c-9271-c7d8f9b1101a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.329410] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5430fdfe-83e5-4a40-acff-0bcc56cd126b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.343278] env[61986]: DEBUG nova.compute.provider_tree [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.374739] env[61986]: DEBUG oslo_vmware.api [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160541, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127824} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.374988] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.375192] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1143.375371] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1143.375547] env[61986]: INFO nova.compute.manager [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1143.375786] env[61986]: DEBUG oslo.service.loopingcall [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.375973] env[61986]: DEBUG nova.compute.manager [-] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1143.376085] env[61986]: DEBUG nova.network.neutron [-] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1143.639400] env[61986]: DEBUG nova.compute.manager [req-b2a28c00-45be-4a95-8191-8ede37f02223 req-005fb51e-e53a-4f47-83be-63972d02da38 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Received event network-vif-deleted-9741a942-9250-4623-a520-3bf6d16b4f69 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.639639] env[61986]: INFO nova.compute.manager [req-b2a28c00-45be-4a95-8191-8ede37f02223 req-005fb51e-e53a-4f47-83be-63972d02da38 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Neutron deleted interface 9741a942-9250-4623-a520-3bf6d16b4f69; detaching it from the instance and deleting it from the info cache [ 1143.639782] env[61986]: DEBUG nova.network.neutron [req-b2a28c00-45be-4a95-8191-8ede37f02223 req-005fb51e-e53a-4f47-83be-63972d02da38 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.845985] env[61986]: DEBUG nova.scheduler.client.report [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.118042] env[61986]: DEBUG nova.network.neutron [-] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.142894] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74c0d644-ecb5-4fa3-bb2c-56346da1d7e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.153503] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c64b238-9f30-4335-adea-cdd584656f27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.183022] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1144.185479] env[61986]: DEBUG nova.compute.manager [req-b2a28c00-45be-4a95-8191-8ede37f02223 req-005fb51e-e53a-4f47-83be-63972d02da38 service nova] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Detach interface failed, port_id=9741a942-9250-4623-a520-3bf6d16b4f69, reason: Instance 43cb9710-3f7d-400f-b046-16c0c77e26b7 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1144.210317] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1144.210597] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1144.210758] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1144.210940] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1144.211141] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1144.211314] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1144.211530] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1144.211693] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1144.211864] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1144.212044] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1144.212256] env[61986]: DEBUG nova.virt.hardware [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.213127] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585d782b-23ca-452f-9074-bb81f9e710c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.220733] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59411bad-b3fa-4cd3-985d-4cb5fc411328 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.350825] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.351450] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1144.353994] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.282s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.355370] env[61986]: INFO nova.compute.claims [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1144.452074] env[61986]: DEBUG nova.network.neutron [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Successfully updated port: 37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1144.620543] env[61986]: INFO nova.compute.manager [-] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Took 1.24 seconds to deallocate network for instance. [ 1144.863438] env[61986]: DEBUG nova.compute.utils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.866925] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1144.867104] env[61986]: DEBUG nova.network.neutron [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1144.913302] env[61986]: DEBUG nova.policy [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c804063142764cac8244fd3d6bd71e16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838620f936dc4489be8b99ef87bf37ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1144.955404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.955557] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.955705] env[61986]: DEBUG nova.network.neutron [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1145.127690] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.160881] env[61986]: DEBUG nova.network.neutron [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Successfully created port: 3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1145.367721] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1145.490889] env[61986]: DEBUG nova.network.neutron [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1145.494789] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f58b2f7-daba-4c2a-ba1d-aea8334d9431 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.502009] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb0e00d-eb7d-4167-b86e-028a9ea66a7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.534362] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dcdc552-32d2-4e77-8600-bee753cf0e68 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.542111] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02aa80fd-311e-44ba-8677-4f0cc501cc21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.555168] env[61986]: DEBUG nova.compute.provider_tree [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.649517] env[61986]: DEBUG nova.network.neutron [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.663662] env[61986]: DEBUG nova.compute.manager [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-vif-plugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.663912] env[61986]: DEBUG oslo_concurrency.lockutils [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.664127] env[61986]: DEBUG oslo_concurrency.lockutils [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.664276] env[61986]: DEBUG oslo_concurrency.lockutils [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.664446] env[61986]: DEBUG nova.compute.manager [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] No waiting events found dispatching network-vif-plugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1145.664611] env[61986]: WARNING nova.compute.manager [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received unexpected event network-vif-plugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 for instance with vm_state building and task_state spawning. [ 1145.664772] env[61986]: DEBUG nova.compute.manager [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.664924] env[61986]: DEBUG nova.compute.manager [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing instance network info cache due to event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1145.665103] env[61986]: DEBUG oslo_concurrency.lockutils [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.060064] env[61986]: DEBUG nova.scheduler.client.report [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.152035] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.152396] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Instance network_info: |[{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1146.152707] env[61986]: DEBUG oslo_concurrency.lockutils [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.152902] env[61986]: DEBUG nova.network.neutron [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1146.154101] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:80:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37976b99-2eee-45d3-93da-ce6f9f9813e2', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1146.161966] env[61986]: DEBUG oslo.service.loopingcall [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.164793] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1146.165260] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-236e3f63-de5c-4690-bbad-4161b374a335 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.185328] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1146.185328] env[61986]: value = "task-1160542" [ 1146.185328] env[61986]: _type = "Task" [ 1146.185328] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.193337] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160542, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.374142] env[61986]: DEBUG nova.network.neutron [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updated VIF entry in instance network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1146.374501] env[61986]: DEBUG nova.network.neutron [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.379869] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1146.406863] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.407139] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.407302] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.407541] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.407711] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.407862] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.408079] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.408248] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.408415] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.408575] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.408751] env[61986]: DEBUG nova.virt.hardware [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.409782] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37991e63-7367-4b54-bcd0-d5c9163abe50 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.417945] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38524c39-f1e3-460a-b2e9-1947d873ded7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.565951] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.212s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.566523] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1146.569712] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.442s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.569928] env[61986]: DEBUG nova.objects.instance [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'resources' on Instance uuid 43cb9710-3f7d-400f-b046-16c0c77e26b7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.623429] env[61986]: DEBUG nova.network.neutron [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Successfully updated port: 3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1146.695673] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160542, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.877569] env[61986]: DEBUG oslo_concurrency.lockutils [req-bde93257-3251-4a9d-a14a-32a368051704 req-31ce2255-d1bf-4866-9b32-cfbe8b5b2865 service nova] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.073153] env[61986]: DEBUG nova.compute.utils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1147.077785] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1147.077785] env[61986]: DEBUG nova.network.neutron [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1147.115080] env[61986]: DEBUG nova.policy [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1147.126369] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "refresh_cache-52353614-d3bb-46da-a8fc-ab920442c64d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.126500] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "refresh_cache-52353614-d3bb-46da-a8fc-ab920442c64d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.126639] env[61986]: DEBUG nova.network.neutron [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1147.196118] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160542, 'name': CreateVM_Task, 'duration_secs': 0.559129} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.198272] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1147.199085] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.199335] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.199620] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1147.199868] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65ffbb16-966f-4db3-b1f6-af16fffea1e3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.205469] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1147.205469] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d4948b-39cf-a369-4c2f-647dce6a18dc" [ 1147.205469] env[61986]: _type = "Task" [ 1147.205469] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.209678] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050a33a7-db67-421b-96c3-0e49b6d1be2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.214840] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d4948b-39cf-a369-4c2f-647dce6a18dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.218970] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db7d664-5eda-4b98-a934-18765d2cc4d7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.249452] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c1dcd0-4071-49f1-9fc6-bdc9c847ca07 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.257198] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22ff67e-686b-4307-9991-47355e8a5407 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.984728] env[61986]: DEBUG nova.network.neutron [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Successfully created port: 7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1147.987928] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1147.993158] env[61986]: DEBUG nova.compute.manager [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Received event network-vif-plugged-3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1147.993392] env[61986]: DEBUG oslo_concurrency.lockutils [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] Acquiring lock "52353614-d3bb-46da-a8fc-ab920442c64d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.993601] env[61986]: DEBUG oslo_concurrency.lockutils [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] Lock "52353614-d3bb-46da-a8fc-ab920442c64d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.993765] env[61986]: DEBUG oslo_concurrency.lockutils [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] Lock "52353614-d3bb-46da-a8fc-ab920442c64d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.993932] env[61986]: DEBUG nova.compute.manager [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] No waiting events found dispatching network-vif-plugged-3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1147.994111] env[61986]: WARNING nova.compute.manager [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Received unexpected event network-vif-plugged-3cce0a97-ab60-4eb5-8c6b-31881dcac86d for instance with vm_state building and task_state spawning. [ 1147.994270] env[61986]: DEBUG nova.compute.manager [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Received event network-changed-3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1147.994424] env[61986]: DEBUG nova.compute.manager [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Refreshing instance network info cache due to event network-changed-3cce0a97-ab60-4eb5-8c6b-31881dcac86d. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1147.994585] env[61986]: DEBUG oslo_concurrency.lockutils [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] Acquiring lock "refresh_cache-52353614-d3bb-46da-a8fc-ab920442c64d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.008979] env[61986]: DEBUG nova.compute.provider_tree [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.013638] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52d4948b-39cf-a369-4c2f-647dce6a18dc, 'name': SearchDatastore_Task, 'duration_secs': 0.009778} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.014166] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.014409] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1148.014636] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.014786] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.014964] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1148.015237] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-267a736c-a67f-421d-88fa-eadedb669242 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.023681] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1148.023876] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1148.025079] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-501e8a1b-0ce3-412c-a18f-10d4e178c461 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.030639] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1148.030639] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e4fe8b-4fc5-efbb-fca1-9e843c233c29" [ 1148.030639] env[61986]: _type = "Task" [ 1148.030639] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.038664] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e4fe8b-4fc5-efbb-fca1-9e843c233c29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.051239] env[61986]: DEBUG nova.network.neutron [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1148.213677] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1148.213958] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1148.286222] env[61986]: DEBUG nova.network.neutron [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Updating instance_info_cache with network_info: [{"id": "3cce0a97-ab60-4eb5-8c6b-31881dcac86d", "address": "fa:16:3e:f0:fa:ff", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cce0a97-ab", "ovs_interfaceid": "3cce0a97-ab60-4eb5-8c6b-31881dcac86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.515174] env[61986]: DEBUG nova.scheduler.client.report [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1148.541342] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52e4fe8b-4fc5-efbb-fca1-9e843c233c29, 'name': SearchDatastore_Task, 'duration_secs': 0.007863} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.542120] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a346e931-0ecb-4eb8-930b-fa10e7b8e779 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.547295] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1148.547295] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bee0fa-feb9-1e3d-888c-4f087e6e89f6" [ 1148.547295] env[61986]: _type = "Task" [ 1148.547295] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.555120] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bee0fa-feb9-1e3d-888c-4f087e6e89f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.789251] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "refresh_cache-52353614-d3bb-46da-a8fc-ab920442c64d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.789628] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Instance network_info: |[{"id": "3cce0a97-ab60-4eb5-8c6b-31881dcac86d", "address": "fa:16:3e:f0:fa:ff", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cce0a97-ab", "ovs_interfaceid": "3cce0a97-ab60-4eb5-8c6b-31881dcac86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1148.789974] env[61986]: DEBUG oslo_concurrency.lockutils [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] Acquired lock "refresh_cache-52353614-d3bb-46da-a8fc-ab920442c64d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.790184] env[61986]: DEBUG nova.network.neutron [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Refreshing network info cache for port 3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1148.791612] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:fa:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cce0a97-ab60-4eb5-8c6b-31881dcac86d', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1148.798939] env[61986]: DEBUG oslo.service.loopingcall [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1148.802133] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1148.802669] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3eb69cac-eaa5-403c-9518-76be72d40edc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.823546] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1148.823546] env[61986]: value = "task-1160543" [ 1148.823546] env[61986]: _type = "Task" [ 1148.823546] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.831545] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160543, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.001517] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1149.004687] env[61986]: DEBUG nova.network.neutron [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Updated VIF entry in instance network info cache for port 3cce0a97-ab60-4eb5-8c6b-31881dcac86d. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1149.005041] env[61986]: DEBUG nova.network.neutron [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Updating instance_info_cache with network_info: [{"id": "3cce0a97-ab60-4eb5-8c6b-31881dcac86d", "address": "fa:16:3e:f0:fa:ff", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cce0a97-ab", "ovs_interfaceid": "3cce0a97-ab60-4eb5-8c6b-31881dcac86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.022690] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.453s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.032853] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1149.033136] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1149.033334] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1149.033483] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1149.033625] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1149.033773] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1149.033973] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1149.034149] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1149.034315] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1149.034475] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1149.034642] env[61986]: DEBUG nova.virt.hardware [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.036012] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8d9059-ac71-4d76-ac07-782a92b59cdc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.043953] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b376ec90-c65e-43fe-865b-054cde08fbf8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.048174] env[61986]: INFO nova.scheduler.client.report [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocations for instance 43cb9710-3f7d-400f-b046-16c0c77e26b7 [ 1149.067456] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52bee0fa-feb9-1e3d-888c-4f087e6e89f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009622} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.068961] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.068961] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1149.068961] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18d22c27-6347-4feb-ba73-1b6a33b56dd9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.074474] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1149.074474] env[61986]: value = "task-1160544" [ 1149.074474] env[61986]: _type = "Task" [ 1149.074474] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.082512] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.336147] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160543, 'name': CreateVM_Task, 'duration_secs': 0.300845} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.336614] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1149.337044] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.337223] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.337555] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1149.337821] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15f7e2d0-e600-41b2-95b5-82bb4a0bac02 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.343433] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1149.343433] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528bc28c-690d-436d-8667-5655339f146d" [ 1149.343433] env[61986]: _type = "Task" [ 1149.343433] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.354094] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528bc28c-690d-436d-8667-5655339f146d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.508484] env[61986]: DEBUG oslo_concurrency.lockutils [req-68692319-3c7b-4c17-85be-b0f25f71ff13 req-e313958a-0db8-4bf0-8dcd-464ae712526c service nova] Releasing lock "refresh_cache-52353614-d3bb-46da-a8fc-ab920442c64d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.566575] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b014db78-d676-4f01-9c2d-215c85256497 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "43cb9710-3f7d-400f-b046-16c0c77e26b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.837s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.583874] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436563} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.584692] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1149.584911] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1149.585188] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad11c699-59d9-45f5-9da8-94e91182b8e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.592268] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1149.592268] env[61986]: value = "task-1160545" [ 1149.592268] env[61986]: _type = "Task" [ 1149.592268] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.600042] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.854598] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]528bc28c-690d-436d-8667-5655339f146d, 'name': SearchDatastore_Task, 'duration_secs': 0.047467} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.854844] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.855085] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1149.855323] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.855470] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.855648] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1149.855897] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-474180d4-852b-4a6d-943b-226f9dc6ff00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.864799] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1149.864978] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1149.865655] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-186a7147-6513-4b88-a97d-31d18bf1a15a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.870373] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1149.870373] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52343ea1-789d-1da0-b827-fef0540c4131" [ 1149.870373] env[61986]: _type = "Task" [ 1149.870373] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.877678] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52343ea1-789d-1da0-b827-fef0540c4131, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.102658] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08056} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.102658] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1150.103611] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71bc3fb-209f-4fac-a1b7-553369c91507 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.125904] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.126220] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df5a057e-e673-4dab-a697-7a98a3de0d1d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.145646] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1150.145646] env[61986]: value = "task-1160546" [ 1150.145646] env[61986]: _type = "Task" [ 1150.145646] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.153338] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.214047] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.214301] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.381905] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52343ea1-789d-1da0-b827-fef0540c4131, 'name': SearchDatastore_Task, 'duration_secs': 0.011389} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.382738] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f15415a1-40fd-4c7c-ae17-15abc3ad349d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.388346] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1150.388346] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521ce316-4fef-7ffe-db2b-0915865ab344" [ 1150.388346] env[61986]: _type = "Task" [ 1150.388346] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.396208] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521ce316-4fef-7ffe-db2b-0915865ab344, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.655771] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160546, 'name': ReconfigVM_Task, 'duration_secs': 0.271206} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.656302] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Reconfigured VM instance instance-0000006d to attach disk [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.657158] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9cf9074a-41b3-4bd2-bf4f-e05c49bd7918 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.663231] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1150.663231] env[61986]: value = "task-1160547" [ 1150.663231] env[61986]: _type = "Task" [ 1150.663231] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.670984] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160547, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.717192] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.717427] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.717665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.717834] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1150.718711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3bc916-1053-44a9-a342-c964f7cd7bed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.726305] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a13448-c463-47b9-a5d1-9850324ae41f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.739673] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e157d80b-eea5-4ac0-b2b2-d68ee3256efd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.746028] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d3a024-e7ff-4480-9664-f2325033199d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.775335] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180762MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1150.775500] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.775695] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.899048] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521ce316-4fef-7ffe-db2b-0915865ab344, 'name': SearchDatastore_Task, 'duration_secs': 0.01475} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.899343] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.899601] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1150.899857] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f7fc190-1e29-4bae-8d3d-c90bdc107706 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.905899] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1150.905899] env[61986]: value = "task-1160548" [ 1150.905899] env[61986]: _type = "Task" [ 1150.905899] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.913169] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160548, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.984072] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.984411] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.174829] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160547, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.416154] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160548, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424648} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.416640] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1151.416918] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1151.417215] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a32f1a3-b5de-4f14-b510-eff8959d7fab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.423900] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1151.423900] env[61986]: value = "task-1160549" [ 1151.423900] env[61986]: _type = "Task" [ 1151.423900] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.434407] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.486975] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1151.673646] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160547, 'name': Rename_Task, 'duration_secs': 0.82501} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.673925] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1151.674181] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-defc4c21-ae1f-4bc0-ab9d-097c7d685fb2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.679976] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1151.679976] env[61986]: value = "task-1160550" [ 1151.679976] env[61986]: _type = "Task" [ 1151.679976] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.687072] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a6094cc4-7d17-4858-abae-972425241c8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 39ea2c99-8b36-457e-96ef-6310223b0d7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 02950912-c04e-4631-ad06-7f57495f00ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance c5a4fd56-25e2-4a95-8d2a-58361ef65e86 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 52353614-d3bb-46da-a8fc-ab920442c64d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.810297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 0250e945-6290-422c-840b-f7afac85331f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1151.868394] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.868705] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.934369] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066834} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.934981] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1151.935809] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79c5164-7002-4c05-95c1-b69a1d611ecf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.958626] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1151.958974] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb65a071-77e5-4947-8993-944c6f170dc3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.979342] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1151.979342] env[61986]: value = "task-1160551" [ 1151.979342] env[61986]: _type = "Task" [ 1151.979342] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.987769] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160551, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.008656] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.025764] env[61986]: DEBUG nova.compute.manager [req-2ce936e6-307d-4d97-8fe6-b706a9326a06 req-1547e542-6f37-4484-89cd-333073958c4c service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Received event network-vif-plugged-7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1152.025993] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ce936e6-307d-4d97-8fe6-b706a9326a06 req-1547e542-6f37-4484-89cd-333073958c4c service nova] Acquiring lock "0250e945-6290-422c-840b-f7afac85331f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.026252] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ce936e6-307d-4d97-8fe6-b706a9326a06 req-1547e542-6f37-4484-89cd-333073958c4c service nova] Lock "0250e945-6290-422c-840b-f7afac85331f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.026387] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ce936e6-307d-4d97-8fe6-b706a9326a06 req-1547e542-6f37-4484-89cd-333073958c4c service nova] Lock "0250e945-6290-422c-840b-f7afac85331f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.026559] env[61986]: DEBUG nova.compute.manager [req-2ce936e6-307d-4d97-8fe6-b706a9326a06 req-1547e542-6f37-4484-89cd-333073958c4c service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] No waiting events found dispatching network-vif-plugged-7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1152.026723] env[61986]: WARNING nova.compute.manager [req-2ce936e6-307d-4d97-8fe6-b706a9326a06 req-1547e542-6f37-4484-89cd-333073958c4c service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Received unexpected event network-vif-plugged-7dba9049-bd65-4709-8aa5-fe7823c36733 for instance with vm_state building and task_state spawning. [ 1152.113841] env[61986]: DEBUG nova.network.neutron [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Successfully updated port: 7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1152.189147] env[61986]: DEBUG oslo_vmware.api [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160550, 'name': PowerOnVM_Task, 'duration_secs': 0.45759} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.189419] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1152.189625] env[61986]: INFO nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Took 8.01 seconds to spawn the instance on the hypervisor. [ 1152.189818] env[61986]: DEBUG nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.190576] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ff5d88-5ae8-4266-b38d-0e30d48dff56 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.312342] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1152.312626] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1152.312796] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1152.371919] env[61986]: DEBUG nova.compute.utils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1152.419947] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e97a57a-ad3f-43d7-b0a4-29b5c4760ce7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.427405] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9590f1dd-b532-4181-ada2-3aaf6ba37f9e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.459165] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334636bc-5a7b-437f-8d1a-91d23d51c669 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.466817] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a12536-fc16-4948-959c-fb85903ab34e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.479788] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.488287] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160551, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.619299] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-0250e945-6290-422c-840b-f7afac85331f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.619467] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-0250e945-6290-422c-840b-f7afac85331f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.619598] env[61986]: DEBUG nova.network.neutron [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.707903] env[61986]: INFO nova.compute.manager [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Took 12.75 seconds to build instance. [ 1152.875349] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.985673] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.994342] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160551, 'name': ReconfigVM_Task, 'duration_secs': 0.685862} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.994625] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1152.995235] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8cafe3e1-2a26-478f-8cf3-beeefc445a34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.001497] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1153.001497] env[61986]: value = "task-1160552" [ 1153.001497] env[61986]: _type = "Task" [ 1153.001497] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.010352] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160552, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.151589] env[61986]: DEBUG nova.network.neutron [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1153.210315] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5abb909c-2c60-4456-88b9-509df0e956d8 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.255s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.280116] env[61986]: DEBUG nova.network.neutron [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Updating instance_info_cache with network_info: [{"id": "7dba9049-bd65-4709-8aa5-fe7823c36733", "address": "fa:16:3e:53:d8:d5", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba9049-bd", "ovs_interfaceid": "7dba9049-bd65-4709-8aa5-fe7823c36733", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.444414] env[61986]: DEBUG nova.compute.manager [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1153.444651] env[61986]: DEBUG nova.compute.manager [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing instance network info cache due to event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1153.444862] env[61986]: DEBUG oslo_concurrency.lockutils [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.445011] env[61986]: DEBUG oslo_concurrency.lockutils [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.445189] env[61986]: DEBUG nova.network.neutron [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1153.489858] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1153.490080] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.714s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.490596] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.482s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.491910] env[61986]: INFO nova.compute.claims [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1153.494369] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.494540] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Cleaning up deleted instances {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1153.512040] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160552, 'name': Rename_Task, 'duration_secs': 0.150917} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.512040] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1153.512245] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dfbdad1b-8753-4a30-9e4a-a7858aae788b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.519400] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1153.519400] env[61986]: value = "task-1160553" [ 1153.519400] env[61986]: _type = "Task" [ 1153.519400] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.533698] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.783911] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-0250e945-6290-422c-840b-f7afac85331f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.783911] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Instance network_info: |[{"id": "7dba9049-bd65-4709-8aa5-fe7823c36733", "address": "fa:16:3e:53:d8:d5", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba9049-bd", "ovs_interfaceid": "7dba9049-bd65-4709-8aa5-fe7823c36733", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1153.784250] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:d8:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7dba9049-bd65-4709-8aa5-fe7823c36733', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1153.792461] env[61986]: DEBUG oslo.service.loopingcall [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.792701] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0250e945-6290-422c-840b-f7afac85331f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1153.792923] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-459ecae3-eac3-43b9-933d-8813fba16a0d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.812040] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1153.812040] env[61986]: value = "task-1160554" [ 1153.812040] env[61986]: _type = "Task" [ 1153.812040] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.822056] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160554, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.951769] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.952045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.952320] env[61986]: INFO nova.compute.manager [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Attaching volume 5b814d91-d6b6-4ea7-9035-1ca3182d197a to /dev/sdb [ 1153.992076] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369c1bca-b020-46ba-b937-2e057d694c10 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.014606] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] There are 63 instances to clean {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1154.014606] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 43cb9710-3f7d-400f-b046-16c0c77e26b7] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.020073] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457f2ac9-4965-4483-818a-8ac6a56e33a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.034928] env[61986]: DEBUG oslo_vmware.api [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160553, 'name': PowerOnVM_Task, 'duration_secs': 0.494157} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.040421] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1154.040581] env[61986]: INFO nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Took 7.66 seconds to spawn the instance on the hypervisor. [ 1154.040770] env[61986]: DEBUG nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.041968] env[61986]: DEBUG nova.virt.block_device [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updating existing volume attachment record: eaac48c6-628e-404f-b92e-7af4198e47af {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1154.045305] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f7b8c4-996b-4800-a8ad-dd2e28e337a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.058622] env[61986]: DEBUG nova.compute.manager [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Received event network-changed-7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1154.058822] env[61986]: DEBUG nova.compute.manager [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Refreshing instance network info cache due to event network-changed-7dba9049-bd65-4709-8aa5-fe7823c36733. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1154.059046] env[61986]: DEBUG oslo_concurrency.lockutils [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] Acquiring lock "refresh_cache-0250e945-6290-422c-840b-f7afac85331f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.059224] env[61986]: DEBUG oslo_concurrency.lockutils [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] Acquired lock "refresh_cache-0250e945-6290-422c-840b-f7afac85331f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.059400] env[61986]: DEBUG nova.network.neutron [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Refreshing network info cache for port 7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1154.204868] env[61986]: DEBUG nova.network.neutron [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updated VIF entry in instance network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.205060] env[61986]: DEBUG nova.network.neutron [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.322219] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160554, 'name': CreateVM_Task, 'duration_secs': 0.319384} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.322445] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0250e945-6290-422c-840b-f7afac85331f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1154.323233] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.323466] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.323849] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1154.324157] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab295f0d-1114-485a-8d58-ec09ee95a18e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.328555] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1154.328555] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5236d30e-90b2-8ede-acd1-1875747c460b" [ 1154.328555] env[61986]: _type = "Task" [ 1154.328555] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.338068] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5236d30e-90b2-8ede-acd1-1875747c460b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.521748] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: fd1b8b0e-d8ff-4912-a5ae-1b8f488728ff] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1154.571013] env[61986]: INFO nova.compute.manager [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Took 13.42 seconds to build instance. [ 1154.652960] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f12283-88c8-4f1a-8fba-6440ad940575 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.662873] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d63f63b-3424-4739-913d-95bbbacf7b24 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.694969] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55453489-693f-44c1-98d0-9c4d3a0ecb2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.703284] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9588c21-0f27-4a8b-8068-88d52ff613d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.708031] env[61986]: DEBUG oslo_concurrency.lockutils [req-066c7a45-fe96-49e3-9880-c568d2646c2d req-5c492727-44f7-4f00-b5e0-3714698ad1a3 service nova] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.716075] env[61986]: DEBUG nova.compute.provider_tree [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.817904] env[61986]: DEBUG nova.network.neutron [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Updated VIF entry in instance network info cache for port 7dba9049-bd65-4709-8aa5-fe7823c36733. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.818313] env[61986]: DEBUG nova.network.neutron [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Updating instance_info_cache with network_info: [{"id": "7dba9049-bd65-4709-8aa5-fe7823c36733", "address": "fa:16:3e:53:d8:d5", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba9049-bd", "ovs_interfaceid": "7dba9049-bd65-4709-8aa5-fe7823c36733", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.840149] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5236d30e-90b2-8ede-acd1-1875747c460b, 'name': SearchDatastore_Task, 'duration_secs': 0.011404} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.840470] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.840708] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1154.840902] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.841064] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.841333] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1154.841610] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba568cc3-b8a6-4c58-9f0b-b7cf76b7d460 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.849988] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1154.850258] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1154.851342] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c26b774-f085-4478-9c09-1423f6521789 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.856361] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1154.856361] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252574c-683a-0be3-d310-91ca7619139f" [ 1154.856361] env[61986]: _type = "Task" [ 1154.856361] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.867240] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252574c-683a-0be3-d310-91ca7619139f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.025402] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 785dc1af-bec0-447b-90f5-2cb1f53bb0f0] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.073397] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4f466bcd-d700-42cc-b519-62e9286a9163 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.927s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.219593] env[61986]: DEBUG nova.scheduler.client.report [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.320843] env[61986]: DEBUG oslo_concurrency.lockutils [req-ae8a72cd-b13a-42a5-ac2f-becea4825bfb req-8e2e572e-119c-4b4a-9412-b8bd6745e354 service nova] Releasing lock "refresh_cache-0250e945-6290-422c-840b-f7afac85331f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.367303] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252574c-683a-0be3-d310-91ca7619139f, 'name': SearchDatastore_Task, 'duration_secs': 0.008292} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.368089] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79c6c85b-a028-4e04-ad10-4c9944b2c80c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.373283] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1155.373283] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b4b154-8dff-db1c-ff87-bcb895ffde7a" [ 1155.373283] env[61986]: _type = "Task" [ 1155.373283] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.380792] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b4b154-8dff-db1c-ff87-bcb895ffde7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.528935] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 18f87ccd-ca34-44f6-aa7d-d38397cc479b] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1155.725673] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.726238] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1155.845734] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "52353614-d3bb-46da-a8fc-ab920442c64d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.845734] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.845734] env[61986]: DEBUG nova.compute.manager [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1155.846412] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede7f976-7a63-4b6a-9749-49cfce08f3b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.852879] env[61986]: DEBUG nova.compute.manager [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61986) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1155.853521] env[61986]: DEBUG nova.objects.instance [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'flavor' on Instance uuid 52353614-d3bb-46da-a8fc-ab920442c64d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.883273] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b4b154-8dff-db1c-ff87-bcb895ffde7a, 'name': SearchDatastore_Task, 'duration_secs': 0.008398} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.883546] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.883793] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 0250e945-6290-422c-840b-f7afac85331f/0250e945-6290-422c-840b-f7afac85331f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1155.884048] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38539947-878f-4e79-bcf5-6fb502083919 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.891040] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1155.891040] env[61986]: value = "task-1160556" [ 1155.891040] env[61986]: _type = "Task" [ 1155.891040] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.897879] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.032818] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: c3bcd1ba-a3df-4b81-9743-431235536642] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.230995] env[61986]: DEBUG nova.compute.utils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.233029] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1156.233029] env[61986]: DEBUG nova.network.neutron [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1156.273260] env[61986]: DEBUG nova.policy [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c40c9363dc7741a3bd6a040f20284837', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e0f7d6b6be498eaec797aa64aaea7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1156.358829] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1156.358829] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f6a25be-c8fa-4a00-bb78-080d0304efcc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.365512] env[61986]: DEBUG oslo_vmware.api [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1156.365512] env[61986]: value = "task-1160558" [ 1156.365512] env[61986]: _type = "Task" [ 1156.365512] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.373949] env[61986]: DEBUG oslo_vmware.api [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.401795] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430808} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.402057] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 0250e945-6290-422c-840b-f7afac85331f/0250e945-6290-422c-840b-f7afac85331f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1156.402298] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1156.402554] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2501f7c8-185a-4d55-9d91-80c200f87595 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.408821] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1156.408821] env[61986]: value = "task-1160559" [ 1156.408821] env[61986]: _type = "Task" [ 1156.408821] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.416465] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.535923] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 9efa4214-3954-49ef-bc7b-6daba8ba07f2] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1156.555276] env[61986]: DEBUG nova.network.neutron [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Successfully created port: 3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1156.735949] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1156.875490] env[61986]: DEBUG oslo_vmware.api [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160558, 'name': PowerOffVM_Task, 'duration_secs': 0.272243} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.875766] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1156.875943] env[61986]: DEBUG nova.compute.manager [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1156.876720] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bb9465-49eb-4c3c-b253-ec59ba728483 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.918419] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06728} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.918678] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.919464] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79edadd-69f8-4323-93e8-c5948c0af01b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.940963] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 0250e945-6290-422c-840b-f7afac85331f/0250e945-6290-422c-840b-f7afac85331f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.941306] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e9b26ec-7bf7-48a7-b1e5-00c54ad14429 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.960319] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1156.960319] env[61986]: value = "task-1160560" [ 1156.960319] env[61986]: _type = "Task" [ 1156.960319] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.968157] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160560, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.040446] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: fc3ae819-efa0-497e-8d2f-d17a53726056] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.387736] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ac5b9e1a-3f6a-4f85-a8ab-bf99c765d969 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.470593] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160560, 'name': ReconfigVM_Task, 'duration_secs': 0.260982} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.470882] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 0250e945-6290-422c-840b-f7afac85331f/0250e945-6290-422c-840b-f7afac85331f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1157.471642] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0f06dfb-4e33-439c-bfe9-52044d150bdb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.478805] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1157.478805] env[61986]: value = "task-1160561" [ 1157.478805] env[61986]: _type = "Task" [ 1157.478805] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.487318] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160561, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.544197] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: e825bf09-7f6c-45db-b6cf-9f2dabb98677] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1157.745795] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1157.782057] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1157.782281] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1157.782411] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.782598] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1157.782751] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.782901] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1157.783131] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1157.783528] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1157.783528] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1157.783685] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1157.783871] env[61986]: DEBUG nova.virt.hardware [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1157.784793] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de0bf76-985c-405b-a18d-2f4007115565 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.793134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d5067b-f2c3-43df-bc2d-3792b59b5d92 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.989626] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160561, 'name': Rename_Task, 'duration_secs': 0.225869} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.989905] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1157.990178] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05290cfa-65e7-4791-80ed-940977fc53af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.997465] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1157.997465] env[61986]: value = "task-1160562" [ 1157.997465] env[61986]: _type = "Task" [ 1157.997465] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.012248] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160562, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.046598] env[61986]: DEBUG nova.network.neutron [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Successfully updated port: 3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1158.048681] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 18893690-92d4-4391-a395-5cf2115397c3] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.055444] env[61986]: DEBUG nova.compute.manager [req-ce9a3e15-d070-40a6-857b-59c1ed7df235 req-1679f131-9444-4115-8e06-8014d8294f6d service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Received event network-vif-plugged-3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1158.055991] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce9a3e15-d070-40a6-857b-59c1ed7df235 req-1679f131-9444-4115-8e06-8014d8294f6d service nova] Acquiring lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.056311] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce9a3e15-d070-40a6-857b-59c1ed7df235 req-1679f131-9444-4115-8e06-8014d8294f6d service nova] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.056605] env[61986]: DEBUG oslo_concurrency.lockutils [req-ce9a3e15-d070-40a6-857b-59c1ed7df235 req-1679f131-9444-4115-8e06-8014d8294f6d service nova] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.057230] env[61986]: DEBUG nova.compute.manager [req-ce9a3e15-d070-40a6-857b-59c1ed7df235 req-1679f131-9444-4115-8e06-8014d8294f6d service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] No waiting events found dispatching network-vif-plugged-3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.057548] env[61986]: WARNING nova.compute.manager [req-ce9a3e15-d070-40a6-857b-59c1ed7df235 req-1679f131-9444-4115-8e06-8014d8294f6d service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Received unexpected event network-vif-plugged-3143b97f-f6d2-4b01-b69c-10dbd22fa45a for instance with vm_state building and task_state spawning. [ 1158.395302] env[61986]: INFO nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Rebuilding instance [ 1158.508180] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160562, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.525085] env[61986]: DEBUG nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.525834] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9cc82a-edf8-4a55-ba13-cd7ee2794fee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.547764] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.548143] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.548143] env[61986]: DEBUG nova.network.neutron [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1158.552505] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: af31ba6f-0a3e-4a1c-af5e-617e8843839f] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1158.599329] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1158.599592] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252486', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'name': 'volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39ea2c99-8b36-457e-96ef-6310223b0d7e', 'attached_at': '', 'detached_at': '', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'serial': '5b814d91-d6b6-4ea7-9035-1ca3182d197a'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1158.600531] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5965eaf-4fd4-4889-a34a-e50f31d1cfe0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.617627] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5126612-7690-40d5-a5f7-7c9fddb668b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.644009] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a/volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.644374] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81234731-769a-421b-8912-9cddbd85a137 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.663021] env[61986]: DEBUG oslo_vmware.api [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1158.663021] env[61986]: value = "task-1160563" [ 1158.663021] env[61986]: _type = "Task" [ 1158.663021] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.671870] env[61986]: DEBUG oslo_vmware.api [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160563, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.007934] env[61986]: DEBUG oslo_vmware.api [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160562, 'name': PowerOnVM_Task, 'duration_secs': 0.579043} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.008228] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1159.008435] env[61986]: INFO nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Took 10.01 seconds to spawn the instance on the hypervisor. [ 1159.008615] env[61986]: DEBUG nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1159.009444] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b41b145-4dcc-4784-81d7-c1a9b9620705 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.036640] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1159.036928] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24aabf67-02f7-4d4e-974c-04728e6a742e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.044057] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1159.044057] env[61986]: value = "task-1160564" [ 1159.044057] env[61986]: _type = "Task" [ 1159.044057] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.052315] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.057014] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 68626872-0875-4ff1-81ad-85bdff30ac94] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.089832] env[61986]: DEBUG nova.network.neutron [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1159.172658] env[61986]: DEBUG oslo_vmware.api [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160563, 'name': ReconfigVM_Task, 'duration_secs': 0.417905} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.172937] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a/volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1159.177560] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-959384c2-b068-4a66-b369-5bfee78cba0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.192994] env[61986]: DEBUG oslo_vmware.api [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1159.192994] env[61986]: value = "task-1160565" [ 1159.192994] env[61986]: _type = "Task" [ 1159.192994] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.201975] env[61986]: DEBUG oslo_vmware.api [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160565, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.235968] env[61986]: DEBUG nova.network.neutron [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Updating instance_info_cache with network_info: [{"id": "3143b97f-f6d2-4b01-b69c-10dbd22fa45a", "address": "fa:16:3e:ce:06:04", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3143b97f-f6", "ovs_interfaceid": "3143b97f-f6d2-4b01-b69c-10dbd22fa45a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.525639] env[61986]: INFO nova.compute.manager [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Took 16.47 seconds to build instance. [ 1159.556824] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1159.557134] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1159.557830] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38beaee6-af0a-4f6e-a70f-24938aa95386 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.560700] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 645ecfee-3c0b-47c1-86c1-8ca21781438a] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1159.567860] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1159.568129] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf82d853-17f8-4734-b250-ce7a30fed0ce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.699258] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1159.699477] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1159.699664] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore1] 52353614-d3bb-46da-a8fc-ab920442c64d {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1159.702682] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0b0d1b8-adbc-4cbf-87d6-a11c1e2fb77a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.704398] env[61986]: DEBUG oslo_vmware.api [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160565, 'name': ReconfigVM_Task, 'duration_secs': 0.17201} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.704674] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252486', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'name': 'volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39ea2c99-8b36-457e-96ef-6310223b0d7e', 'attached_at': '', 'detached_at': '', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'serial': '5b814d91-d6b6-4ea7-9035-1ca3182d197a'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1159.710943] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1159.710943] env[61986]: value = "task-1160567" [ 1159.710943] env[61986]: _type = "Task" [ 1159.710943] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.718489] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160567, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.739282] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.739653] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Instance network_info: |[{"id": "3143b97f-f6d2-4b01-b69c-10dbd22fa45a", "address": "fa:16:3e:ce:06:04", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3143b97f-f6", "ovs_interfaceid": "3143b97f-f6d2-4b01-b69c-10dbd22fa45a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1159.740085] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:06:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3143b97f-f6d2-4b01-b69c-10dbd22fa45a', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1159.747679] env[61986]: DEBUG oslo.service.loopingcall [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.747894] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1159.748417] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1e36ec1-cbbe-4542-8057-ac59a4c79f82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.767702] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1159.767702] env[61986]: value = "task-1160568" [ 1159.767702] env[61986]: _type = "Task" [ 1159.767702] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.775738] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160568, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.027920] env[61986]: DEBUG oslo_concurrency.lockutils [None req-4a81e7e2-9626-4d05-9bb8-64fb191ea76f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "0250e945-6290-422c-840b-f7afac85331f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.981s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.064622] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: db9050b8-897e-4d62-8dc2-3b334ab8afd4] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.082802] env[61986]: DEBUG nova.compute.manager [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Received event network-changed-3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1160.082990] env[61986]: DEBUG nova.compute.manager [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Refreshing instance network info cache due to event network-changed-3143b97f-f6d2-4b01-b69c-10dbd22fa45a. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1160.083229] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] Acquiring lock "refresh_cache-a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.083593] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] Acquired lock "refresh_cache-a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.083593] env[61986]: DEBUG nova.network.neutron [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Refreshing network info cache for port 3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1160.116698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "0250e945-6290-422c-840b-f7afac85331f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.116949] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "0250e945-6290-422c-840b-f7afac85331f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.117200] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "0250e945-6290-422c-840b-f7afac85331f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.117398] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "0250e945-6290-422c-840b-f7afac85331f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.117568] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "0250e945-6290-422c-840b-f7afac85331f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.119996] env[61986]: INFO nova.compute.manager [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Terminating instance [ 1160.121863] env[61986]: DEBUG nova.compute.manager [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1160.122065] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1160.122966] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d7b19d-1d1e-4e89-87bd-8e933b9ff682 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.130452] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1160.130681] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13581129-08dc-479b-84b8-6b252d7e5da3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.136201] env[61986]: DEBUG oslo_vmware.api [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1160.136201] env[61986]: value = "task-1160569" [ 1160.136201] env[61986]: _type = "Task" [ 1160.136201] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.143934] env[61986]: DEBUG oslo_vmware.api [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.221640] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160567, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142764} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.221774] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1160.221964] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1160.222179] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1160.277550] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160568, 'name': CreateVM_Task, 'duration_secs': 0.316475} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.277738] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1160.278455] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.278605] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.278920] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1160.279197] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6e03f8b-1ee2-4592-b1a7-f4c2c3c20690 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.283753] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1160.283753] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521f06aa-6e8c-6fdb-13bc-0765fb9fd84f" [ 1160.283753] env[61986]: _type = "Task" [ 1160.283753] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.292055] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521f06aa-6e8c-6fdb-13bc-0765fb9fd84f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.568293] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: eb10ea87-8019-4f3b-af3f-32a901ec6a07] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1160.645980] env[61986]: DEBUG oslo_vmware.api [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160569, 'name': PowerOffVM_Task, 'duration_secs': 0.222564} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.646277] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1160.646454] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1160.646701] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e45c6dda-f940-4138-b7a9-226da66bd608 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.714503] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1160.714773] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1160.714999] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore1] 0250e945-6290-422c-840b-f7afac85331f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1160.715300] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b61dbc8d-e791-4cc7-a26a-2dfda2a50342 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.722371] env[61986]: DEBUG oslo_vmware.api [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1160.722371] env[61986]: value = "task-1160571" [ 1160.722371] env[61986]: _type = "Task" [ 1160.722371] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.735208] env[61986]: DEBUG oslo_vmware.api [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.747270] env[61986]: DEBUG nova.objects.instance [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'flavor' on Instance uuid 39ea2c99-8b36-457e-96ef-6310223b0d7e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.795711] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521f06aa-6e8c-6fdb-13bc-0765fb9fd84f, 'name': SearchDatastore_Task, 'duration_secs': 0.008147} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.796594] env[61986]: DEBUG nova.network.neutron [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Updated VIF entry in instance network info cache for port 3143b97f-f6d2-4b01-b69c-10dbd22fa45a. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1160.796936] env[61986]: DEBUG nova.network.neutron [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Updating instance_info_cache with network_info: [{"id": "3143b97f-f6d2-4b01-b69c-10dbd22fa45a", "address": "fa:16:3e:ce:06:04", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3143b97f-f6", "ovs_interfaceid": "3143b97f-f6d2-4b01-b69c-10dbd22fa45a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.798191] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.798440] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.798678] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.798825] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.798998] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.799444] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bbcf61a-013a-4af0-8131-a83449b2c875 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.807324] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.807444] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1160.808583] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c132a0d6-589c-4864-af30-8c340f7f98af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.813723] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1160.813723] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5231fedf-6302-8f53-2ec7-f0533b39c9a3" [ 1160.813723] env[61986]: _type = "Task" [ 1160.813723] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.822435] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5231fedf-6302-8f53-2ec7-f0533b39c9a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.071467] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 8a594fe5-6fff-48ab-9f7f-474b2a24a486] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.235151] env[61986]: DEBUG oslo_vmware.api [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183323} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.235614] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1161.235799] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1161.235978] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1161.236174] env[61986]: INFO nova.compute.manager [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 0250e945-6290-422c-840b-f7afac85331f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1161.236485] env[61986]: DEBUG oslo.service.loopingcall [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.236660] env[61986]: DEBUG nova.compute.manager [-] [instance: 0250e945-6290-422c-840b-f7afac85331f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1161.236731] env[61986]: DEBUG nova.network.neutron [-] [instance: 0250e945-6290-422c-840b-f7afac85331f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1161.254033] env[61986]: DEBUG oslo_concurrency.lockutils [None req-de398629-ca31-4232-bd5b-fba0d38e7890 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.302s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.260668] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1161.260912] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1161.261187] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1161.261300] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1161.261446] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1161.261599] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1161.261808] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1161.261970] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1161.262160] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1161.262326] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1161.262505] env[61986]: DEBUG nova.virt.hardware [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1161.263405] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49c405a-5c2e-4581-aecd-c4bfa8ae1e40 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.272067] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e85dd06-27c4-42eb-bbbe-001f63ed76eb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.286015] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:fa:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cce0a97-ab60-4eb5-8c6b-31881dcac86d', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1161.293473] env[61986]: DEBUG oslo.service.loopingcall [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.295917] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1161.296168] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94d68223-4b94-4ddf-a18d-f13bf82b2fae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.310386] env[61986]: DEBUG oslo_concurrency.lockutils [req-2ab14d43-2694-4e4e-acaa-c7cf24721d0c req-e1eaa755-16de-405c-8f91-affce87101bd service nova] Releasing lock "refresh_cache-a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.319397] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1161.319397] env[61986]: value = "task-1160572" [ 1161.319397] env[61986]: _type = "Task" [ 1161.319397] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.325858] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5231fedf-6302-8f53-2ec7-f0533b39c9a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008294} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.326868] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5a94ef5-7ae9-4247-90dc-f1a87f312158 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.331582] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160572, 'name': CreateVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.334356] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1161.334356] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520dbd3c-5e4a-21bd-8d83-74aed1ae4723" [ 1161.334356] env[61986]: _type = "Task" [ 1161.334356] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.342961] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520dbd3c-5e4a-21bd-8d83-74aed1ae4723, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.410607] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.410897] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.574150] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 45d80bab-0140-4473-9dc2-4a5481838aed] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1161.829171] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160572, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.843340] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520dbd3c-5e4a-21bd-8d83-74aed1ae4723, 'name': SearchDatastore_Task, 'duration_secs': 0.009164} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.843579] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.843824] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c/a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1161.844073] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec22d123-892c-4b86-8b7d-22de00b0ef62 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.849620] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1161.849620] env[61986]: value = "task-1160573" [ 1161.849620] env[61986]: _type = "Task" [ 1161.849620] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.856627] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.914407] env[61986]: INFO nova.compute.manager [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Detaching volume 5b814d91-d6b6-4ea7-9035-1ca3182d197a [ 1161.953661] env[61986]: INFO nova.virt.block_device [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Attempting to driver detach volume 5b814d91-d6b6-4ea7-9035-1ca3182d197a from mountpoint /dev/sdb [ 1161.953937] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1161.954158] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252486', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'name': 'volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39ea2c99-8b36-457e-96ef-6310223b0d7e', 'attached_at': '', 'detached_at': '', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'serial': '5b814d91-d6b6-4ea7-9035-1ca3182d197a'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1161.955102] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea5b57a-c643-4cf1-b183-3cc845251dd3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.976872] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efbc8b6-771b-45dc-9d3c-7399f372de86 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.983468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101dc675-77f0-477f-ba39-d3d46d9a032c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.004526] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb100d1-814d-4969-bc1a-b74fcd7c252f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.021547] env[61986]: DEBUG nova.network.neutron [-] [instance: 0250e945-6290-422c-840b-f7afac85331f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.022973] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] The volume has not been displaced from its original location: [datastore2] volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a/volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1162.028939] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1162.029616] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8110db23-5413-4f45-b757-aba9c2d3bc5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.051111] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1162.051111] env[61986]: value = "task-1160574" [ 1162.051111] env[61986]: _type = "Task" [ 1162.051111] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.062526] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160574, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.077654] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 73b2c463-0460-4a68-b788-ab0c7818d8f8] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.111177] env[61986]: DEBUG nova.compute.manager [req-0fb82f0b-55d3-49f6-91bd-aa1b604f52f5 req-b934c7dc-59f7-47e1-b2be-1e381f0c3db7 service nova] [instance: 0250e945-6290-422c-840b-f7afac85331f] Received event network-vif-deleted-7dba9049-bd65-4709-8aa5-fe7823c36733 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1162.329785] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160572, 'name': CreateVM_Task} progress is 99%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.357749] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160573, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453349} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.358108] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c/a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1162.358377] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.358634] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-600811a6-2c49-4603-91b6-c69fe2b06ad2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.364465] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1162.364465] env[61986]: value = "task-1160575" [ 1162.364465] env[61986]: _type = "Task" [ 1162.364465] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.372270] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.531102] env[61986]: INFO nova.compute.manager [-] [instance: 0250e945-6290-422c-840b-f7afac85331f] Took 1.29 seconds to deallocate network for instance. [ 1162.560965] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.581515] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 92077300-2e56-4dec-9e8d-9b566ac3fe0e] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1162.829958] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160572, 'name': CreateVM_Task, 'duration_secs': 1.376824} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.830168] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1162.830834] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.831017] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.831398] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1162.831645] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10111fba-964c-4ca3-98c6-e46ebb704f4b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.835644] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1162.835644] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f2fa68-2dc6-e12f-e985-45f803efb75d" [ 1162.835644] env[61986]: _type = "Task" [ 1162.835644] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.842505] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f2fa68-2dc6-e12f-e985-45f803efb75d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.871705] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057786} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.871932] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1162.872676] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5092929a-ad92-4f6b-bce2-fb3bf0edba7e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.894569] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c/a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.894790] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-190d5ac8-b503-4c73-a242-4f506966609c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.913031] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1162.913031] env[61986]: value = "task-1160576" [ 1162.913031] env[61986]: _type = "Task" [ 1162.913031] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.920229] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160576, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.039763] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.040115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.040395] env[61986]: DEBUG nova.objects.instance [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid 0250e945-6290-422c-840b-f7afac85331f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.061066] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160574, 'name': ReconfigVM_Task, 'duration_secs': 0.741233} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.061333] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1163.066193] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88e0ab3b-9115-4eb9-b278-a0f3c6b52f69 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.081646] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1163.081646] env[61986]: value = "task-1160577" [ 1163.081646] env[61986]: _type = "Task" [ 1163.081646] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.086713] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bda7750f-eec9-40d9-ace5-18d48234126e] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.093160] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.345928] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52f2fa68-2dc6-e12f-e985-45f803efb75d, 'name': SearchDatastore_Task, 'duration_secs': 0.008149} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.346247] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.346479] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1163.346712] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.346860] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.347051] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1163.347299] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74ec66ca-1db6-4245-8e7d-8053e96ad917 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.354982] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1163.355170] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1163.355820] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e4f5fe6-d032-4e9e-88ee-979a248389c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.360709] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1163.360709] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbac7d-57dc-2216-d097-b491580ef2db" [ 1163.360709] env[61986]: _type = "Task" [ 1163.360709] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.367976] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbac7d-57dc-2216-d097-b491580ef2db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.421459] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160576, 'name': ReconfigVM_Task, 'duration_secs': 0.308339} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.421678] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Reconfigured VM instance instance-00000070 to attach disk [datastore2] a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c/a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.422295] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bedc3a99-a885-4692-b328-12bf657f050f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.427539] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1163.427539] env[61986]: value = "task-1160578" [ 1163.427539] env[61986]: _type = "Task" [ 1163.427539] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.434633] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160578, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.589609] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 3b925174-cf0b-4c6c-b6b6-1c99a50026be] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1163.594615] env[61986]: DEBUG oslo_vmware.api [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160577, 'name': ReconfigVM_Task, 'duration_secs': 0.143604} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.597186] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252486', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'name': 'volume-5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '39ea2c99-8b36-457e-96ef-6310223b0d7e', 'attached_at': '', 'detached_at': '', 'volume_id': '5b814d91-d6b6-4ea7-9035-1ca3182d197a', 'serial': '5b814d91-d6b6-4ea7-9035-1ca3182d197a'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1163.675494] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e845350d-151d-441a-8d5e-0aa03d566278 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.683218] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0f6c9e-474a-428a-9eea-d05f4e437b1d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.713850] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6960552e-a1c4-4341-94fd-07b7c7d9951c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.721247] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf90a0f-8823-45b5-8975-bc38a0ea9108 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.735891] env[61986]: DEBUG nova.compute.provider_tree [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.870639] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fbac7d-57dc-2216-d097-b491580ef2db, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.871436] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00e450c3-7aaf-42a2-837a-5b18a5d61a11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.876669] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1163.876669] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274ec7e-9187-83b9-9876-6668a07449f3" [ 1163.876669] env[61986]: _type = "Task" [ 1163.876669] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.883853] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274ec7e-9187-83b9-9876-6668a07449f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.936461] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160578, 'name': Rename_Task, 'duration_secs': 0.132155} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.936675] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1163.936916] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b77d1495-9f16-4a46-a4f6-1871c85821bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.942957] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1163.942957] env[61986]: value = "task-1160579" [ 1163.942957] env[61986]: _type = "Task" [ 1163.942957] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.949922] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.097020] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 221e10a3-da31-410c-80f8-4bcc2c515710] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.141209] env[61986]: DEBUG nova.objects.instance [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'flavor' on Instance uuid 39ea2c99-8b36-457e-96ef-6310223b0d7e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.239434] env[61986]: DEBUG nova.scheduler.client.report [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1164.387849] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5274ec7e-9187-83b9-9876-6668a07449f3, 'name': SearchDatastore_Task, 'duration_secs': 0.009109} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.388130] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.388384] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1164.388634] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0b3599a-bbef-4211-a70b-df1afdcad39e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.394312] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1164.394312] env[61986]: value = "task-1160580" [ 1164.394312] env[61986]: _type = "Task" [ 1164.394312] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.401442] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.451719] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160579, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.599735] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: bc93bdd5-256e-4900-a9ae-609b0a7d402d] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1164.744144] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.765332] env[61986]: INFO nova.scheduler.client.report [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance 0250e945-6290-422c-840b-f7afac85331f [ 1164.905110] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160580, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.956023] env[61986]: DEBUG oslo_vmware.api [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160579, 'name': PowerOnVM_Task, 'duration_secs': 0.669909} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.956023] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1164.956023] env[61986]: INFO nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1164.956023] env[61986]: DEBUG nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1164.956023] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a834c2d5-11a1-4e91-94a4-5993bae00c0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.102997] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f9aa0511-b05c-408f-ac06-c49bf8dc648c] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.150717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-ea85450c-d990-4e74-a9c9-8dec46be6fdc tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.740s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.274463] env[61986]: DEBUG oslo_concurrency.lockutils [None req-1264cfff-d040-4916-8081-509da81597f4 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "0250e945-6290-422c-840b-f7afac85331f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.157s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.405285] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523795} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.405662] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1165.405950] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1165.406250] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0764c5ba-bade-4c32-a87b-a7348bc1f565 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.412049] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1165.412049] env[61986]: value = "task-1160581" [ 1165.412049] env[61986]: _type = "Task" [ 1165.412049] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.419516] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.474952] env[61986]: INFO nova.compute.manager [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Took 13.48 seconds to build instance. [ 1165.606053] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f15bf19d-b86a-4b0a-ac1a-9df8e77e9382] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1165.614297] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "02950912-c04e-4631-ad06-7f57495f00ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.614698] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "02950912-c04e-4631-ad06-7f57495f00ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.615044] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "02950912-c04e-4631-ad06-7f57495f00ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.615366] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "02950912-c04e-4631-ad06-7f57495f00ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.615664] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "02950912-c04e-4631-ad06-7f57495f00ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.618757] env[61986]: INFO nova.compute.manager [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Terminating instance [ 1165.621853] env[61986]: DEBUG nova.compute.manager [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.622605] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1165.623710] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5f97eb-9c7d-4316-a3af-7a9a19a63420 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.634622] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1165.635030] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67e9b3bd-b7fd-4647-8e53-f5636a8c55d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.643735] env[61986]: DEBUG oslo_vmware.api [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1165.643735] env[61986]: value = "task-1160582" [ 1165.643735] env[61986]: _type = "Task" [ 1165.643735] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.655727] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.656147] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.656514] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "39ea2c99-8b36-457e-96ef-6310223b0d7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.656855] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.657187] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.659816] env[61986]: DEBUG oslo_vmware.api [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.660490] env[61986]: INFO nova.compute.manager [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Terminating instance [ 1165.663259] env[61986]: DEBUG nova.compute.manager [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.663590] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1165.664541] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7719a517-4daa-48f6-81e9-b3f72957083f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.672389] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1165.672649] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03c26031-2707-4c3e-8947-1d6c895ccc77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.679406] env[61986]: DEBUG oslo_vmware.api [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1165.679406] env[61986]: value = "task-1160583" [ 1165.679406] env[61986]: _type = "Task" [ 1165.679406] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.687332] env[61986]: DEBUG oslo_vmware.api [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.923479] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059264} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.923850] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1165.924872] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c168e5dc-1221-453c-87e1-baf332086124 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.946755] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.947084] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cf1269a-9108-4d51-9d14-9ebb3e5306f7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.967568] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1165.967568] env[61986]: value = "task-1160584" [ 1165.967568] env[61986]: _type = "Task" [ 1165.967568] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.976793] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e4bcde06-8c27-45eb-b0dd-3ae8688c2d0a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.992s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.976950] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160584, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.110855] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 5170aa51-3307-42b1-b0dd-645dd4036e5b] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.155065] env[61986]: DEBUG oslo_vmware.api [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160582, 'name': PowerOffVM_Task, 'duration_secs': 0.231227} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.155065] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1166.155065] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1166.155065] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de242c53-7798-4dd8-865a-0c44cb3ccd3a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.189687] env[61986]: DEBUG oslo_vmware.api [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160583, 'name': PowerOffVM_Task, 'duration_secs': 0.19888} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.191015] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1166.191273] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1166.193978] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56f48912-4e3d-457f-96c3-2bbd6996a095 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.244472] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1166.244773] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1166.245038] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore2] 02950912-c04e-4631-ad06-7f57495f00ea {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.245355] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97095485-9af2-4065-9e3a-a4a8546fe26e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.251477] env[61986]: DEBUG oslo_vmware.api [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1166.251477] env[61986]: value = "task-1160587" [ 1166.251477] env[61986]: _type = "Task" [ 1166.251477] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.259088] env[61986]: DEBUG oslo_vmware.api [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.262038] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1166.262170] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1166.262324] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleting the datastore file [datastore2] 39ea2c99-8b36-457e-96ef-6310223b0d7e {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.262556] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-954ee3da-ad7c-4ed8-aba1-3036cfe691ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.267906] env[61986]: DEBUG oslo_vmware.api [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for the task: (returnval){ [ 1166.267906] env[61986]: value = "task-1160588" [ 1166.267906] env[61986]: _type = "Task" [ 1166.267906] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.275525] env[61986]: DEBUG oslo_vmware.api [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160588, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.476710] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160584, 'name': ReconfigVM_Task, 'duration_secs': 0.261827} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.477016] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 52353614-d3bb-46da-a8fc-ab920442c64d/52353614-d3bb-46da-a8fc-ab920442c64d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1166.477652] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce4b93da-c5e5-463a-9822-176ec0c49da9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.483058] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1166.483058] env[61986]: value = "task-1160589" [ 1166.483058] env[61986]: _type = "Task" [ 1166.483058] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.491124] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160589, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.614909] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 3d5f4513-bbc4-404e-9d3e-340bd369fc3c] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1166.699421] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1556ee5b-600a-47f7-9d0b-c3edae584025 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.706354] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Suspending the VM {{(pid=61986) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1166.706621] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e4b2b968-80a4-4927-ae80-fea3363dd34f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.713054] env[61986]: DEBUG oslo_vmware.api [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1166.713054] env[61986]: value = "task-1160590" [ 1166.713054] env[61986]: _type = "Task" [ 1166.713054] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.724130] env[61986]: DEBUG oslo_vmware.api [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160590, 'name': SuspendVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.760858] env[61986]: DEBUG oslo_vmware.api [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323935} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.761129] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.761393] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1166.761577] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1166.761741] env[61986]: INFO nova.compute.manager [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1166.761981] env[61986]: DEBUG oslo.service.loopingcall [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.762206] env[61986]: DEBUG nova.compute.manager [-] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1166.762307] env[61986]: DEBUG nova.network.neutron [-] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1166.776134] env[61986]: DEBUG oslo_vmware.api [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Task: {'id': task-1160588, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317104} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.776473] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.776606] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1166.776755] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1166.776925] env[61986]: INFO nova.compute.manager [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1166.777177] env[61986]: DEBUG oslo.service.loopingcall [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.777396] env[61986]: DEBUG nova.compute.manager [-] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1166.777895] env[61986]: DEBUG nova.network.neutron [-] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1166.995225] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160589, 'name': Rename_Task, 'duration_secs': 0.151461} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.995225] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1166.995225] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab2650ba-dbf7-434e-bd29-094a82bd1ded {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.002602] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1167.002602] env[61986]: value = "task-1160591" [ 1167.002602] env[61986]: _type = "Task" [ 1167.002602] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.015180] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160591, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.118391] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 93c4f104-1812-4bb7-bfa7-cbf70a19ff51] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.223427] env[61986]: DEBUG oslo_vmware.api [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160590, 'name': SuspendVM_Task} progress is 62%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.425913] env[61986]: DEBUG nova.compute.manager [req-a805b33b-60ca-4773-be72-f2f072df2f78 req-78c092e8-a6a8-49ba-86be-84162cbc1c30 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Received event network-vif-deleted-2219b83b-06dd-449f-9274-be9d72376018 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1167.426170] env[61986]: INFO nova.compute.manager [req-a805b33b-60ca-4773-be72-f2f072df2f78 req-78c092e8-a6a8-49ba-86be-84162cbc1c30 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Neutron deleted interface 2219b83b-06dd-449f-9274-be9d72376018; detaching it from the instance and deleting it from the info cache [ 1167.426811] env[61986]: DEBUG nova.network.neutron [req-a805b33b-60ca-4773-be72-f2f072df2f78 req-78c092e8-a6a8-49ba-86be-84162cbc1c30 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.514831] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160591, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.623056] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: dea128a9-aed0-40b2-ae17-c068ea8e3452] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1167.725757] env[61986]: DEBUG oslo_vmware.api [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160590, 'name': SuspendVM_Task, 'duration_secs': 0.790405} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.726136] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Suspended the VM {{(pid=61986) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1167.726434] env[61986]: DEBUG nova.compute.manager [None req-5303144c-a235-4af9-a132-44b634dbad4a tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1167.727417] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14f4920-40bc-41d6-85cc-97b8df25a8be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.903689] env[61986]: DEBUG nova.network.neutron [-] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.929594] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84d5352d-45c6-450e-bc30-1da57098ed21 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.939392] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3a335f-130d-411d-bdee-05ea78d0e557 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.967431] env[61986]: DEBUG nova.compute.manager [req-a805b33b-60ca-4773-be72-f2f072df2f78 req-78c092e8-a6a8-49ba-86be-84162cbc1c30 service nova] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Detach interface failed, port_id=2219b83b-06dd-449f-9274-be9d72376018, reason: Instance 02950912-c04e-4631-ad06-7f57495f00ea could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1168.012779] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160591, 'name': PowerOnVM_Task, 'duration_secs': 0.619225} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.012984] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1168.013212] env[61986]: DEBUG nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1168.014049] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb3b515-414c-4f2a-801f-58456618cbe2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.083110] env[61986]: DEBUG nova.network.neutron [-] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.124902] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 49bcac13-6dde-4a28-9131-f9a1e1d2e386] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.406197] env[61986]: INFO nova.compute.manager [-] [instance: 02950912-c04e-4631-ad06-7f57495f00ea] Took 1.64 seconds to deallocate network for instance. [ 1168.524744] env[61986]: INFO nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] bringing vm to original state: 'stopped' [ 1168.587449] env[61986]: INFO nova.compute.manager [-] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Took 1.81 seconds to deallocate network for instance. [ 1168.628521] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6488f68e-7b84-4462-aef5-25d02db504f6] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1168.913123] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.913413] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.913672] env[61986]: DEBUG nova.objects.instance [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid 02950912-c04e-4631-ad06-7f57495f00ea {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.092435] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.132551] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 59cfee64-ef79-4b8f-a703-c8812551d12f] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.138975] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.139296] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.140505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.140505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.140505] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.141909] env[61986]: INFO nova.compute.manager [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Terminating instance [ 1169.143644] env[61986]: DEBUG nova.compute.manager [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1169.143848] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1169.145020] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9095db3d-dc70-4b25-a4e3-f5376787c2b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.152565] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1169.152790] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3985909-a081-4f3f-b46e-3da8074e7cc3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.213411] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1169.213770] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1169.214039] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleting the datastore file [datastore2] a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1169.214321] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89c1af14-58f8-4f29-bdff-607e5c64b184 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.220660] env[61986]: DEBUG oslo_vmware.api [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1169.220660] env[61986]: value = "task-1160593" [ 1169.220660] env[61986]: _type = "Task" [ 1169.220660] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.228367] env[61986]: DEBUG oslo_vmware.api [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160593, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.453356] env[61986]: DEBUG nova.compute.manager [req-eec0466b-c745-4497-89ab-93452c880060 req-4776b75b-bce3-4a08-b35d-c29142dc7278 service nova] [instance: 39ea2c99-8b36-457e-96ef-6310223b0d7e] Received event network-vif-deleted-dba5f768-392c-41ae-9284-b299120172b9 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1169.522230] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34af32e4-afe5-4c7d-8b7a-d6201b7bdcb9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.529183] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce37b6c0-99a5-47fb-bc49-dda46251bdc3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.532506] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "52353614-d3bb-46da-a8fc-ab920442c64d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.532717] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.532894] env[61986]: DEBUG nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1169.533601] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480088e2-fc22-4428-af74-66bd9b62388b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.539230] env[61986]: DEBUG nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61986) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1169.563829] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1169.564536] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2850b4-51b2-4aa5-ad9f-f5148c965511 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.567009] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28e7a8c5-ebd6-4375-9dba-a19c5b06ca80 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.574573] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1977f7b-aca5-4185-82cd-a7133c946010 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.578033] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1169.578033] env[61986]: value = "task-1160594" [ 1169.578033] env[61986]: _type = "Task" [ 1169.578033] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.588550] env[61986]: DEBUG nova.compute.provider_tree [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.595220] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.636227] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 3512cd72-6666-4810-828e-50230956c4b1] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1169.731264] env[61986]: DEBUG oslo_vmware.api [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160593, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139078} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.731543] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1169.731746] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1169.731948] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1169.732173] env[61986]: INFO nova.compute.manager [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1169.732473] env[61986]: DEBUG oslo.service.loopingcall [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1169.732697] env[61986]: DEBUG nova.compute.manager [-] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1169.732855] env[61986]: DEBUG nova.network.neutron [-] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1170.087516] env[61986]: DEBUG oslo_vmware.api [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160594, 'name': PowerOffVM_Task, 'duration_secs': 0.170069} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.087794] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1170.088227] env[61986]: DEBUG nova.compute.manager [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1170.088784] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e996eb0e-6004-4123-963a-3bfb2bc113df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.091800] env[61986]: DEBUG nova.scheduler.client.report [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1170.139777] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: f42e92ef-8a90-473c-8662-57cbb9e3f4ee] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.475538] env[61986]: DEBUG nova.network.neutron [-] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.596827] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.599206] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.507s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.599450] env[61986]: DEBUG nova.objects.instance [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lazy-loading 'resources' on Instance uuid 39ea2c99-8b36-457e-96ef-6310223b0d7e {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.604719] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.072s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.625060] env[61986]: INFO nova.scheduler.client.report [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance 02950912-c04e-4631-ad06-7f57495f00ea [ 1170.643316] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: b833adc9-2967-4b0b-81c2-0b8deac20f69] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1170.703537] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.703537] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.703537] env[61986]: INFO nova.compute.manager [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Shelving [ 1170.977613] env[61986]: INFO nova.compute.manager [-] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Took 1.24 seconds to deallocate network for instance. [ 1171.115202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.132948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-81d2ee9e-ac6e-409b-a9d5-7457354b179e tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "02950912-c04e-4631-ad06-7f57495f00ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.518s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.146669] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6f9349b7-886a-4077-8f6d-a9800ab353fe] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.187209] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf81572-fafd-4550-9989-7ce9512106f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.195394] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bcda560-c0e0-4bb8-ba7f-c692f00ac6d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.228973] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9890d49-4701-4f07-8a3b-a7427bd28eaa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.231610] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1171.231872] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85b5807d-3c51-4a85-b8aa-0859b88fc30f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.239245] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c82ddf-2da0-4710-943b-184b5a3dfc77 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.242989] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1171.242989] env[61986]: value = "task-1160595" [ 1171.242989] env[61986]: _type = "Task" [ 1171.242989] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.253991] env[61986]: DEBUG nova.compute.provider_tree [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.260091] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.397642] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "52353614-d3bb-46da-a8fc-ab920442c64d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.397953] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.398202] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "52353614-d3bb-46da-a8fc-ab920442c64d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.398490] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.398668] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.401130] env[61986]: INFO nova.compute.manager [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Terminating instance [ 1171.403500] env[61986]: DEBUG nova.compute.manager [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1171.403703] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1171.404539] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fea318-9196-4ecd-9e52-cb84c4e6b29a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.411681] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1171.411904] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47aa8494-1962-4895-a779-70f768c60be5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.473947] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1171.474271] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1171.474518] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore2] 52353614-d3bb-46da-a8fc-ab920442c64d {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1171.474845] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8ef1efb-e403-405e-a170-72ae1afe2296 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.482262] env[61986]: DEBUG nova.compute.manager [req-a18b1468-ea86-44ec-8ad9-3fe0ee95f7a4 req-6344669d-1407-4074-bc90-f56d256c7c9e service nova] [instance: a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c] Received event network-vif-deleted-3143b97f-f6d2-4b01-b69c-10dbd22fa45a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1171.484439] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.484747] env[61986]: DEBUG oslo_vmware.api [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1171.484747] env[61986]: value = "task-1160597" [ 1171.484747] env[61986]: _type = "Task" [ 1171.484747] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.492766] env[61986]: DEBUG oslo_vmware.api [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160597, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.650073] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: a101dc3c-ca6b-4a72-a9b4-051b077a10fd] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1171.708230] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.708467] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.752827] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160595, 'name': PowerOffVM_Task, 'duration_secs': 0.1669} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.753107] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1171.753866] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d53121e-60c9-43f1-b338-b9f0db6b5b32 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.756820] env[61986]: DEBUG nova.scheduler.client.report [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.775880] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a82fe17-4830-49c6-a352-3be9052b06b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.993901] env[61986]: DEBUG oslo_vmware.api [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160597, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145795} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.994176] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.994411] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1171.994550] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1171.994727] env[61986]: INFO nova.compute.manager [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1171.994968] env[61986]: DEBUG oslo.service.loopingcall [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.995181] env[61986]: DEBUG nova.compute.manager [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.995279] env[61986]: DEBUG nova.network.neutron [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1172.153152] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: d72cd38b-ef14-467b-bf53-97d9e66534c8] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.211163] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1172.261753] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.662s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.264064] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.149s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.264265] env[61986]: DEBUG nova.objects.instance [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1172.286696] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1172.287858] env[61986]: INFO nova.scheduler.client.report [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Deleted allocations for instance 39ea2c99-8b36-457e-96ef-6310223b0d7e [ 1172.288990] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e27c333c-6361-467b-8b7e-13a1deabd09e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.299869] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1172.299869] env[61986]: value = "task-1160598" [ 1172.299869] env[61986]: _type = "Task" [ 1172.299869] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.309727] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160598, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.658052] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 9b539a94-6c39-46f8-b194-27047245d1f2] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1172.732108] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.742941] env[61986]: DEBUG nova.network.neutron [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.797787] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f532da4e-ece6-4b87-ac40-26233bd4e838 tempest-AttachVolumeNegativeTest-2096245214 tempest-AttachVolumeNegativeTest-2096245214-project-member] Lock "39ea2c99-8b36-457e-96ef-6310223b0d7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.142s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.810514] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160598, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.161714] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 98337bb6-9502-4d4c-af00-028659b246bf] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1173.245937] env[61986]: INFO nova.compute.manager [-] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Took 1.25 seconds to deallocate network for instance. [ 1173.273292] env[61986]: DEBUG oslo_concurrency.lockutils [None req-c4b475f9-8ae7-4f99-af1f-52c3ff3b298e tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.274313] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.790s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.274538] env[61986]: DEBUG nova.objects.instance [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'resources' on Instance uuid a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.310343] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160598, 'name': CreateSnapshot_Task, 'duration_secs': 0.556712} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.310600] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1173.311337] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cba8e4-a46b-4546-88b8-1b6a564e3c93 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.508975] env[61986]: DEBUG nova.compute.manager [req-4aadeabd-a110-48f3-bf22-03768a9a6b11 req-78375a2d-2fc4-4920-8ce1-79e62e101b0f service nova] [instance: 52353614-d3bb-46da-a8fc-ab920442c64d] Received event network-vif-deleted-3cce0a97-ab60-4eb5-8c6b-31881dcac86d {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1173.664671] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 4f13b629-e2a7-4668-9d77-eb638078e246] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1173.752785] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.829134] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1173.831594] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c64caf7f-ac28-406b-9fc5-8a0ccdffe023 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.842143] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1173.842143] env[61986]: value = "task-1160600" [ 1173.842143] env[61986]: _type = "Task" [ 1173.842143] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.854008] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160600, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.883022] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308f7fbe-da4e-4e6f-8480-f74bc874fa13 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.890930] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73899721-bb74-4840-9c76-b21e3e76cfc2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.923088] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccdbb75-a781-456f-b279-63ca91ba2d51 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.931396] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2911aca-6188-4fcd-a10d-9d1afb01cf88 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.945532] env[61986]: DEBUG nova.compute.provider_tree [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.168591] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: b9706e51-0f74-4dbc-aab1-ea640ac78c62] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1174.352984] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160600, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.448852] env[61986]: DEBUG nova.scheduler.client.report [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1174.672541] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 52eda388-05f4-416c-ac9d-bd1c1e31ba9c] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1174.853435] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160600, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.954179] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.956811] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.225s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.958362] env[61986]: INFO nova.compute.claims [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1174.972899] env[61986]: INFO nova.scheduler.client.report [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocations for instance a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c [ 1175.175902] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 076c8e91-af70-489f-89d5-f598166fafc6] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1175.354426] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160600, 'name': CloneVM_Task, 'duration_secs': 1.165698} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.354695] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Created linked-clone VM from snapshot [ 1175.355443] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786a63aa-f499-4b4f-9094-cce3d75543d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.364293] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Uploading image b72c8918-6428-4315-9eb6-2e52bebfc404 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1175.386109] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1175.386109] env[61986]: value = "vm-252490" [ 1175.386109] env[61986]: _type = "VirtualMachine" [ 1175.386109] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1175.386404] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2e99c2c0-ba3c-44a0-a0a2-03bf3bdeaad6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.394180] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease: (returnval){ [ 1175.394180] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fd3763-7c10-459f-3b80-0c5d423b71d0" [ 1175.394180] env[61986]: _type = "HttpNfcLease" [ 1175.394180] env[61986]: } obtained for exporting VM: (result){ [ 1175.394180] env[61986]: value = "vm-252490" [ 1175.394180] env[61986]: _type = "VirtualMachine" [ 1175.394180] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1175.394449] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the lease: (returnval){ [ 1175.394449] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fd3763-7c10-459f-3b80-0c5d423b71d0" [ 1175.394449] env[61986]: _type = "HttpNfcLease" [ 1175.394449] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1175.400918] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1175.400918] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fd3763-7c10-459f-3b80-0c5d423b71d0" [ 1175.400918] env[61986]: _type = "HttpNfcLease" [ 1175.400918] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1175.481027] env[61986]: DEBUG oslo_concurrency.lockutils [None req-df70c7c2-16bc-4317-8b1d-4245cadb9bd3 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "a9ee2b1f-2d15-4d42-aec2-bb8355ec2b2c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.342s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.679742] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 46b18b6c-6e40-45b7-9d3f-6177b08a52a5] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1175.902778] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1175.902778] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fd3763-7c10-459f-3b80-0c5d423b71d0" [ 1175.902778] env[61986]: _type = "HttpNfcLease" [ 1175.902778] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1175.903134] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1175.903134] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52fd3763-7c10-459f-3b80-0c5d423b71d0" [ 1175.903134] env[61986]: _type = "HttpNfcLease" [ 1175.903134] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1175.903888] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4973638d-bfe9-468e-a5eb-1e551e35a828 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.911310] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8cc57-fcc5-b865-fb1f-5f8e44b52b8f/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1175.911495] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8cc57-fcc5-b865-fb1f-5f8e44b52b8f/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1175.999808] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ca1eccc3-2ced-49ae-904a-87cff86b97a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.065694] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a231011-dfcd-4ecd-ad1c-cd1b4fbadf66 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.073787] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2b8bc1-3c59-4639-a8d7-6651105f0f1e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.103155] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba9583c-9b12-4e87-9460-091ba6675661 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.109632] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.109861] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.114157] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c360aad-2d7e-474b-8b9f-5bb987372c94 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.129760] env[61986]: DEBUG nova.compute.provider_tree [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.182610] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: a0fa2cbe-d8c3-462a-9744-aaa36a8d314e] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1176.614042] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1176.632823] env[61986]: DEBUG nova.scheduler.client.report [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.685243] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 021a27ec-cee0-454d-8daf-e6a82bd9330c] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1177.137454] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.138291] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.138995] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1177.141770] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.389s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.142054] env[61986]: DEBUG nova.objects.instance [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'resources' on Instance uuid 52353614-d3bb-46da-a8fc-ab920442c64d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.188504] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6f38006d-b1f5-4aeb-9124-1e32378c22b2] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1177.646107] env[61986]: DEBUG nova.compute.utils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1177.661069] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1177.661324] env[61986]: DEBUG nova.network.neutron [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1177.692019] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 182479b8-f72f-4395-99a4-af0d6f91f7d4] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1177.729341] env[61986]: DEBUG nova.policy [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1177.795582] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6398045a-4cdd-4664-918c-a6c7013f0611 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.804858] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4c01d9-2385-48c2-a743-deda6f1bcddb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.842129] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0f8d5c-5e02-41f0-a140-7e2525375c65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.850601] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33557fdd-c6c6-49ab-bd0b-0760d5aa9f29 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.865446] env[61986]: DEBUG nova.compute.provider_tree [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.014020] env[61986]: DEBUG nova.network.neutron [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Successfully created port: 938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1178.162088] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1178.204664] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: ee47cab9-5dfd-48ce-ba70-cb800d735b19] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.369076] env[61986]: DEBUG nova.scheduler.client.report [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1178.707819] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 92f9d5e4-22a9-4b77-8e1e-8d4d5e48ce0f] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.873976] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.732s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.876539] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.739s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.878219] env[61986]: INFO nova.compute.claims [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1178.901375] env[61986]: INFO nova.scheduler.client.report [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted allocations for instance 52353614-d3bb-46da-a8fc-ab920442c64d [ 1179.172087] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1179.196876] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1179.197151] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1179.197349] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1179.197512] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1179.197830] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1179.198026] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1179.198242] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1179.198408] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1179.198617] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1179.198750] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1179.198927] env[61986]: DEBUG nova.virt.hardware [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1179.199878] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dab5075-e316-4723-971d-973c89c3bad7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.209048] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1328c5ab-4bca-4a91-9744-e40e8f42c4e0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.213250] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 9e57f8f9-6e9e-45fb-91d1-132490e930ae] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1179.409501] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e30630ab-fa11-48c6-a6e9-04d25dfe6fe0 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "52353614-d3bb-46da-a8fc-ab920442c64d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.011s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.464916] env[61986]: DEBUG nova.compute.manager [req-7bfc43fe-b4d7-4c51-ba99-aae40bc15dc4 req-890928ee-a954-47af-99f7-daccdbd20449 service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Received event network-vif-plugged-938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.465241] env[61986]: DEBUG oslo_concurrency.lockutils [req-7bfc43fe-b4d7-4c51-ba99-aae40bc15dc4 req-890928ee-a954-47af-99f7-daccdbd20449 service nova] Acquiring lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.465532] env[61986]: DEBUG oslo_concurrency.lockutils [req-7bfc43fe-b4d7-4c51-ba99-aae40bc15dc4 req-890928ee-a954-47af-99f7-daccdbd20449 service nova] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.465740] env[61986]: DEBUG oslo_concurrency.lockutils [req-7bfc43fe-b4d7-4c51-ba99-aae40bc15dc4 req-890928ee-a954-47af-99f7-daccdbd20449 service nova] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.465970] env[61986]: DEBUG nova.compute.manager [req-7bfc43fe-b4d7-4c51-ba99-aae40bc15dc4 req-890928ee-a954-47af-99f7-daccdbd20449 service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] No waiting events found dispatching network-vif-plugged-938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1179.466189] env[61986]: WARNING nova.compute.manager [req-7bfc43fe-b4d7-4c51-ba99-aae40bc15dc4 req-890928ee-a954-47af-99f7-daccdbd20449 service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Received unexpected event network-vif-plugged-938a568f-0336-4dc6-80f3-2f5994ef2ad2 for instance with vm_state building and task_state spawning. [ 1179.716857] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 86c9e45c-b3f1-4004-9acc-190b11a4a926] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1179.982095] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c392609f-1ff3-4368-8795-bdac5fe0e9fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.988799] env[61986]: DEBUG nova.network.neutron [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Successfully updated port: 938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1179.993394] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed79ab4b-7b83-42c2-af79-bb3ceef898b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.025456] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e008f207-0442-4ea7-a862-633b13e9fdd7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.035054] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49ec449-978d-4cb2-9324-a6d02f4f00ca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.050386] env[61986]: DEBUG nova.compute.provider_tree [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.074479] env[61986]: DEBUG nova.compute.manager [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Received event network-changed-938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1180.074709] env[61986]: DEBUG nova.compute.manager [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Refreshing instance network info cache due to event network-changed-938a568f-0336-4dc6-80f3-2f5994ef2ad2. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1180.074962] env[61986]: DEBUG oslo_concurrency.lockutils [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] Acquiring lock "refresh_cache-d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.075102] env[61986]: DEBUG oslo_concurrency.lockutils [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] Acquired lock "refresh_cache-d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.075266] env[61986]: DEBUG nova.network.neutron [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Refreshing network info cache for port 938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1180.220205] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: ebf5fee5-7f1b-4537-aec3-77a8a963670b] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1180.495912] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.555175] env[61986]: DEBUG nova.scheduler.client.report [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1180.605343] env[61986]: DEBUG nova.network.neutron [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1180.699127] env[61986]: DEBUG nova.network.neutron [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.724118] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6f20dd8e-9289-458d-bed6-f2ef9daaa917] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.060018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.060665] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1181.202621] env[61986]: DEBUG oslo_concurrency.lockutils [req-f2b8860a-7864-4ed1-add1-058f763edb0c req-eb9c626a-59b9-4fee-9857-801240f1089a service nova] Releasing lock "refresh_cache-d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.203106] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.203205] env[61986]: DEBUG nova.network.neutron [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1181.227568] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: eaa61f81-2f9b-4d1c-bab8-4363fb71e936] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.565883] env[61986]: DEBUG nova.compute.utils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1181.567362] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1181.567534] env[61986]: DEBUG nova.network.neutron [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1181.609944] env[61986]: DEBUG nova.policy [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c40c9363dc7741a3bd6a040f20284837', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2e0f7d6b6be498eaec797aa64aaea7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1181.731579] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 8438d2a7-975c-4d1f-8906-f960bfadfe39] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.747966] env[61986]: DEBUG nova.network.neutron [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1181.955880] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.956160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.000586] env[61986]: DEBUG nova.network.neutron [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Updating instance_info_cache with network_info: [{"id": "938a568f-0336-4dc6-80f3-2f5994ef2ad2", "address": "fa:16:3e:41:67:93", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap938a568f-03", "ovs_interfaceid": "938a568f-0336-4dc6-80f3-2f5994ef2ad2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.004905] env[61986]: DEBUG nova.network.neutron [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Successfully created port: 9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1182.071341] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1182.235411] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: ce09db77-1eac-4228-a4f8-228a3b7e69d9] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1182.458991] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1182.503425] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.503855] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Instance network_info: |[{"id": "938a568f-0336-4dc6-80f3-2f5994ef2ad2", "address": "fa:16:3e:41:67:93", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap938a568f-03", "ovs_interfaceid": "938a568f-0336-4dc6-80f3-2f5994ef2ad2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1182.504731] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:67:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '938a568f-0336-4dc6-80f3-2f5994ef2ad2', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1182.514198] env[61986]: DEBUG oslo.service.loopingcall [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1182.514611] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1182.514877] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba645b04-0a2b-4bfd-ab99-97a110eda710 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.537957] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1182.537957] env[61986]: value = "task-1160604" [ 1182.537957] env[61986]: _type = "Task" [ 1182.537957] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.547331] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160604, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.738835] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: c1a6ffac-99bd-492c-99e5-f7c46b352d8e] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1182.983871] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.984204] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.985945] env[61986]: INFO nova.compute.claims [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1183.048079] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160604, 'name': CreateVM_Task, 'duration_secs': 0.36728} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.048289] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1183.049087] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.049302] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.049688] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1183.049988] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4199a1f-5bf2-4b9e-a979-2266a0bb3f97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.055906] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1183.055906] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52408d37-dbf1-da13-2554-2e0154ab0357" [ 1183.055906] env[61986]: _type = "Task" [ 1183.055906] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.065087] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52408d37-dbf1-da13-2554-2e0154ab0357, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.081467] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1183.107304] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1183.107581] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1183.107799] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1183.107917] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1183.108083] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1183.108243] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1183.108459] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1183.108623] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1183.108794] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1183.108959] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1183.109154] env[61986]: DEBUG nova.virt.hardware [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1183.110135] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd57dc68-9157-41e9-9901-9fbfa6e779a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.118894] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdec797c-8e21-4771-9232-77b41bef666d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.246582] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 977799b4-2793-4513-9447-483146fc7ac4] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1183.454638] env[61986]: DEBUG nova.compute.manager [req-a6090c46-207d-48a7-92fe-d04f5ef1484c req-9c0f95d8-148a-4df9-b452-840214679143 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Received event network-vif-plugged-9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1183.454826] env[61986]: DEBUG oslo_concurrency.lockutils [req-a6090c46-207d-48a7-92fe-d04f5ef1484c req-9c0f95d8-148a-4df9-b452-840214679143 service nova] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.455346] env[61986]: DEBUG oslo_concurrency.lockutils [req-a6090c46-207d-48a7-92fe-d04f5ef1484c req-9c0f95d8-148a-4df9-b452-840214679143 service nova] Lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.455612] env[61986]: DEBUG oslo_concurrency.lockutils [req-a6090c46-207d-48a7-92fe-d04f5ef1484c req-9c0f95d8-148a-4df9-b452-840214679143 service nova] Lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.455798] env[61986]: DEBUG nova.compute.manager [req-a6090c46-207d-48a7-92fe-d04f5ef1484c req-9c0f95d8-148a-4df9-b452-840214679143 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] No waiting events found dispatching network-vif-plugged-9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1183.455972] env[61986]: WARNING nova.compute.manager [req-a6090c46-207d-48a7-92fe-d04f5ef1484c req-9c0f95d8-148a-4df9-b452-840214679143 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Received unexpected event network-vif-plugged-9f357ac4-99e2-44ce-8f2c-3de7619c93fa for instance with vm_state building and task_state spawning. [ 1183.558115] env[61986]: DEBUG nova.network.neutron [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Successfully updated port: 9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1183.571016] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52408d37-dbf1-da13-2554-2e0154ab0357, 'name': SearchDatastore_Task, 'duration_secs': 0.01265} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.571463] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.571709] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1183.571961] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.572131] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.572344] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1183.572623] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff8dd701-b5ab-4ca0-ac86-be6ea98fc205 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.583044] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1183.583223] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1183.584021] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4f16451-6959-4513-8177-3eeee5d57f82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.590814] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1183.590814] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52121fc2-0729-3709-834f-446142ed1aac" [ 1183.590814] env[61986]: _type = "Task" [ 1183.590814] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.599459] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52121fc2-0729-3709-834f-446142ed1aac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.750069] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 6e0697d4-d862-497c-9bdd-dd1ef2d4272b] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1183.882017] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8cc57-fcc5-b865-fb1f-5f8e44b52b8f/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1183.882980] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c812692c-cd2f-4b82-a658-e0ff60adf609 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.890585] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8cc57-fcc5-b865-fb1f-5f8e44b52b8f/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1183.890768] env[61986]: ERROR oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8cc57-fcc5-b865-fb1f-5f8e44b52b8f/disk-0.vmdk due to incomplete transfer. [ 1183.891013] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-80d0f2e6-5e98-49c3-8ac1-6131afbcf5a4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.898997] env[61986]: DEBUG oslo_vmware.rw_handles [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8cc57-fcc5-b865-fb1f-5f8e44b52b8f/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1183.899269] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Uploaded image b72c8918-6428-4315-9eb6-2e52bebfc404 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1183.901610] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1183.901877] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a7a24c62-d3d4-42e4-8299-4a86abad34ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.909258] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1183.909258] env[61986]: value = "task-1160605" [ 1183.909258] env[61986]: _type = "Task" [ 1183.909258] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.919436] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160605, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.066335] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.066503] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.066668] env[61986]: DEBUG nova.network.neutron [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1184.101553] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52121fc2-0729-3709-834f-446142ed1aac, 'name': SearchDatastore_Task, 'duration_secs': 0.013147} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.103422] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73775470-9f6f-4c92-91da-e7f293aff245 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.105730] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d1a4c73-294c-4573-a212-d0e01ab6abe3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.111020] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1184.111020] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527bf00d-8913-da3b-e3c2-2bb4660e0ae7" [ 1184.111020] env[61986]: _type = "Task" [ 1184.111020] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.116100] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c6ef04-3ac4-41b6-b192-e52d2116fa7f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.123987] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]527bf00d-8913-da3b-e3c2-2bb4660e0ae7, 'name': SearchDatastore_Task, 'duration_secs': 0.009888} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.148776] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.149084] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] d818b2ef-5c36-4df8-a3e0-1d48e027dd9d/d818b2ef-5c36-4df8-a3e0-1d48e027dd9d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1184.149937] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f76c00a3-631e-4304-830e-d9fa9feaa1d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.152481] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d488c568-99db-4dbe-a287-b2767576f35f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.161160] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1d304c-fde3-4536-a91d-ef695f9dc209 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.165759] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1184.165759] env[61986]: value = "task-1160606" [ 1184.165759] env[61986]: _type = "Task" [ 1184.165759] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.177480] env[61986]: DEBUG nova.compute.provider_tree [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.183615] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160606, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.253130] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 40de5f22-0de2-466a-91ab-dcb6ec586dad] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1184.421089] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160605, 'name': Destroy_Task, 'duration_secs': 0.344586} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.421482] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Destroyed the VM [ 1184.421756] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1184.422048] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5b453000-b5f7-407a-a54d-c2526884e511 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.432620] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1184.432620] env[61986]: value = "task-1160607" [ 1184.432620] env[61986]: _type = "Task" [ 1184.432620] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.444455] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160607, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.602648] env[61986]: DEBUG nova.network.neutron [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1184.678447] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160606, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481372} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.678684] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] d818b2ef-5c36-4df8-a3e0-1d48e027dd9d/d818b2ef-5c36-4df8-a3e0-1d48e027dd9d.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1184.678903] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1184.679171] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c52f677-be1d-4ae3-87b8-d06e2ff6dbb3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.681611] env[61986]: DEBUG nova.scheduler.client.report [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1184.689586] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1184.689586] env[61986]: value = "task-1160608" [ 1184.689586] env[61986]: _type = "Task" [ 1184.689586] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.698822] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160608, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.745308] env[61986]: DEBUG nova.network.neutron [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [{"id": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "address": "fa:16:3e:e5:1f:66", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f357ac4-99", "ovs_interfaceid": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.756623] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa800cea-6c7c-48ce-bfc6-a00f2f4cff7d] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1184.943229] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160607, 'name': RemoveSnapshot_Task, 'duration_secs': 0.46444} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.943770] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1184.943844] env[61986]: DEBUG nova.compute.manager [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.944700] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaaf249d-e0f1-4c2b-8fbb-6f2a2b8b600e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.186755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.187325] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1185.199530] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160608, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075702} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.199777] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1185.200544] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b864738-5464-40ac-89a8-659a30e256b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.224090] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] d818b2ef-5c36-4df8-a3e0-1d48e027dd9d/d818b2ef-5c36-4df8-a3e0-1d48e027dd9d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1185.225017] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34216885-1c58-4407-a6a6-10d92e9aca36 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.244586] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1185.244586] env[61986]: value = "task-1160609" [ 1185.244586] env[61986]: _type = "Task" [ 1185.244586] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.247832] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.248117] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Instance network_info: |[{"id": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "address": "fa:16:3e:e5:1f:66", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f357ac4-99", "ovs_interfaceid": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1185.248492] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:1f:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f357ac4-99e2-44ce-8f2c-3de7619c93fa', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1185.255732] env[61986]: DEBUG oslo.service.loopingcall [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1185.256501] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1185.256741] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49aa0d1d-61b7-4094-afd9-76231792d60f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.273739] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: aa12919e-3891-4da9-a280-4155137864e1] Instance has had 0 of 5 cleanup attempts {{(pid=61986) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1185.275607] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.280558] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1185.280558] env[61986]: value = "task-1160610" [ 1185.280558] env[61986]: _type = "Task" [ 1185.280558] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.288179] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160610, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.456958] env[61986]: INFO nova.compute.manager [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Shelve offloading [ 1185.458704] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1185.458954] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b70e7981-f1b6-4000-b444-3d446e36814b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.467722] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1185.467722] env[61986]: value = "task-1160611" [ 1185.467722] env[61986]: _type = "Task" [ 1185.467722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.476164] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160611, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.480294] env[61986]: DEBUG nova.compute.manager [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Received event network-changed-9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1185.480507] env[61986]: DEBUG nova.compute.manager [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Refreshing instance network info cache due to event network-changed-9f357ac4-99e2-44ce-8f2c-3de7619c93fa. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1185.480689] env[61986]: DEBUG oslo_concurrency.lockutils [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] Acquiring lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.480848] env[61986]: DEBUG oslo_concurrency.lockutils [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] Acquired lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.481064] env[61986]: DEBUG nova.network.neutron [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Refreshing network info cache for port 9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1185.696079] env[61986]: DEBUG nova.compute.utils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1185.697547] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1185.697699] env[61986]: DEBUG nova.network.neutron [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1185.735917] env[61986]: DEBUG nova.policy [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c804063142764cac8244fd3d6bd71e16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838620f936dc4489be8b99ef87bf37ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1185.754910] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.777143] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.777524] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Cleaning up deleted instances with incomplete migration {{(pid=61986) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1185.790298] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160610, 'name': CreateVM_Task, 'duration_secs': 0.362578} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.791022] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1185.791703] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.791900] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.792218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1185.792721] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee334369-f08f-44c5-8cf5-0e940a7530d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.798372] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1185.798372] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52934187-b660-4aa0-05ef-df2efcbd77cc" [ 1185.798372] env[61986]: _type = "Task" [ 1185.798372] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.806219] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52934187-b660-4aa0-05ef-df2efcbd77cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.978707] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1185.978707] env[61986]: DEBUG nova.compute.manager [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.979394] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f2f2b2-210a-4a89-bda8-344ca83cba2e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.991246] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.991246] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.991246] env[61986]: DEBUG nova.network.neutron [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1186.040947] env[61986]: DEBUG nova.network.neutron [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Successfully created port: 0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1186.200614] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1186.258479] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160609, 'name': ReconfigVM_Task, 'duration_secs': 0.799133} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.258832] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Reconfigured VM instance instance-00000071 to attach disk [datastore1] d818b2ef-5c36-4df8-a3e0-1d48e027dd9d/d818b2ef-5c36-4df8-a3e0-1d48e027dd9d.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.259747] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7b958b3-1db9-4f61-b668-b2475e6e43f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.271094] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1186.271094] env[61986]: value = "task-1160612" [ 1186.271094] env[61986]: _type = "Task" [ 1186.271094] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.282968] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160612, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.283283] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.309949] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52934187-b660-4aa0-05ef-df2efcbd77cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.310321] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.310580] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1186.310850] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.311252] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.311252] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1186.311487] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd72c0c9-9c17-41c3-a7b4-385f3124f7a8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.320654] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1186.320848] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1186.321623] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5de7c56a-91b6-459d-8931-b9f14e9f7aae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.327428] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1186.327428] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f283e-93d3-b0e1-da32-0fd1edc8179f" [ 1186.327428] env[61986]: _type = "Task" [ 1186.327428] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.332519] env[61986]: DEBUG nova.network.neutron [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updated VIF entry in instance network info cache for port 9f357ac4-99e2-44ce-8f2c-3de7619c93fa. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1186.332959] env[61986]: DEBUG nova.network.neutron [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [{"id": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "address": "fa:16:3e:e5:1f:66", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f357ac4-99", "ovs_interfaceid": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.336735] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f283e-93d3-b0e1-da32-0fd1edc8179f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.712113] env[61986]: DEBUG nova.network.neutron [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.782351] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160612, 'name': Rename_Task, 'duration_secs': 0.146553} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.782649] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1186.782877] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f079781-4bfb-42bb-864b-ae0af4518cca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.790237] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1186.790237] env[61986]: value = "task-1160613" [ 1186.790237] env[61986]: _type = "Task" [ 1186.790237] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.797895] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.838705] env[61986]: DEBUG oslo_concurrency.lockutils [req-8dfbcc91-e8b8-4bf1-a7d3-0ff6e59be5ec req-30d1013b-8087-4a06-8863-ee881bddb455 service nova] Releasing lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.839150] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]525f283e-93d3-b0e1-da32-0fd1edc8179f, 'name': SearchDatastore_Task, 'duration_secs': 0.015349} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.839927] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed50d1e7-1c45-44c9-af2c-4e2552681e78 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.845439] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1186.845439] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5253d2e1-27af-5044-9450-74cedc1f5f69" [ 1186.845439] env[61986]: _type = "Task" [ 1186.845439] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.853753] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5253d2e1-27af-5044-9450-74cedc1f5f69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.215946] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1187.219106] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.245362] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1187.245635] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1187.245794] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1187.245977] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1187.246144] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1187.246298] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1187.246547] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1187.246743] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1187.246919] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1187.247099] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1187.247281] env[61986]: DEBUG nova.virt.hardware [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1187.248163] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3df443-9516-497a-9e85-4f071eb18d61 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.257317] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0638ca67-91cb-4910-a284-5bd5d493d6c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.299752] env[61986]: DEBUG oslo_vmware.api [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160613, 'name': PowerOnVM_Task, 'duration_secs': 0.469594} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.300016] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1187.300227] env[61986]: INFO nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Took 8.13 seconds to spawn the instance on the hypervisor. [ 1187.300418] env[61986]: DEBUG nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1187.301167] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d93239-dbae-4a3e-aef3-afb6eeb93c36 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.356660] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5253d2e1-27af-5044-9450-74cedc1f5f69, 'name': SearchDatastore_Task, 'duration_secs': 0.009247} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.356918] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.357195] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f/cff8354d-1eb9-446c-8c63-6697264c373f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1187.357449] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5825d956-6fc7-4105-a418-88300576da3e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.365334] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1187.365334] env[61986]: value = "task-1160614" [ 1187.365334] env[61986]: _type = "Task" [ 1187.365334] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.375121] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160614, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.512477] env[61986]: DEBUG nova.compute.manager [req-50effdba-ecb5-4347-95d0-22ecfcf79a74 req-25bd1d90-0b90-4306-9600-203683bcd8f9 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-vif-unplugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1187.512735] env[61986]: DEBUG oslo_concurrency.lockutils [req-50effdba-ecb5-4347-95d0-22ecfcf79a74 req-25bd1d90-0b90-4306-9600-203683bcd8f9 service nova] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.512957] env[61986]: DEBUG oslo_concurrency.lockutils [req-50effdba-ecb5-4347-95d0-22ecfcf79a74 req-25bd1d90-0b90-4306-9600-203683bcd8f9 service nova] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.513172] env[61986]: DEBUG oslo_concurrency.lockutils [req-50effdba-ecb5-4347-95d0-22ecfcf79a74 req-25bd1d90-0b90-4306-9600-203683bcd8f9 service nova] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.513441] env[61986]: DEBUG nova.compute.manager [req-50effdba-ecb5-4347-95d0-22ecfcf79a74 req-25bd1d90-0b90-4306-9600-203683bcd8f9 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] No waiting events found dispatching network-vif-unplugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1187.513713] env[61986]: WARNING nova.compute.manager [req-50effdba-ecb5-4347-95d0-22ecfcf79a74 req-25bd1d90-0b90-4306-9600-203683bcd8f9 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received unexpected event network-vif-unplugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f for instance with vm_state shelved and task_state shelving_offloading. [ 1187.677561] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1187.678599] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2157aa5-74cb-48fd-b1af-fced7222cda5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.687571] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1187.687915] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8892a2f-e0f1-489e-bb90-0cea07247620 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.780674] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1187.781179] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1187.781615] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleting the datastore file [datastore2] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.782473] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f062f15-928d-4c2e-a0c0-1f9881fc6da0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.786272] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.786485] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.796223] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1187.796223] env[61986]: value = "task-1160616" [ 1187.796223] env[61986]: _type = "Task" [ 1187.796223] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.802321] env[61986]: DEBUG nova.network.neutron [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Successfully updated port: 0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1187.814737] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160616, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.818671] env[61986]: INFO nova.compute.manager [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Took 15.10 seconds to build instance. [ 1187.875755] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160614, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466171} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.876121] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f/cff8354d-1eb9-446c-8c63-6697264c373f.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1187.876403] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1187.876628] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de8e319f-3f26-4e8a-91ef-9aa108a86757 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.884917] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1187.884917] env[61986]: value = "task-1160617" [ 1187.884917] env[61986]: _type = "Task" [ 1187.884917] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.893400] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.293235] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.293407] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1188.305496] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.305641] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.305788] env[61986]: DEBUG nova.network.neutron [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1188.306889] env[61986]: DEBUG oslo_vmware.api [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160616, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145997} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.308161] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.308161] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1188.308161] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1188.320358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f12ba0a3-da98-402c-8d2b-62d3bba4b7c8 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.612s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.332806] env[61986]: INFO nova.scheduler.client.report [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted allocations for instance 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 [ 1188.395762] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.244822} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.396355] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1188.397164] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf557db7-63ea-4c4d-b97e-351040940f37 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.419653] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f/cff8354d-1eb9-446c-8c63-6697264c373f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.419937] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d263ccb-edbb-484c-9328-371fb98a0134 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.440445] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1188.440445] env[61986]: value = "task-1160618" [ 1188.440445] env[61986]: _type = "Task" [ 1188.440445] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.449732] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.818936] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.819300] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.819564] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.819766] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.819951] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.822348] env[61986]: INFO nova.compute.manager [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Terminating instance [ 1188.824558] env[61986]: DEBUG nova.compute.manager [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1188.824558] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1188.825233] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa4a8e5-b3b2-44c4-b980-07412bd4ff2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.833289] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1188.833491] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32cf96ad-3f52-4f2f-89a3-980e06b59190 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.837045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.837286] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.837584] env[61986]: DEBUG nova.objects.instance [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'resources' on Instance uuid 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.839151] env[61986]: DEBUG nova.network.neutron [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1188.842944] env[61986]: DEBUG oslo_vmware.api [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1188.842944] env[61986]: value = "task-1160619" [ 1188.842944] env[61986]: _type = "Task" [ 1188.842944] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.852657] env[61986]: DEBUG oslo_vmware.api [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.951148] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160618, 'name': ReconfigVM_Task, 'duration_secs': 0.281446} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.951463] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Reconfigured VM instance instance-00000072 to attach disk [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f/cff8354d-1eb9-446c-8c63-6697264c373f.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.952125] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2089d3b2-13d5-48d1-814b-213903813d64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.962874] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1188.962874] env[61986]: value = "task-1160620" [ 1188.962874] env[61986]: _type = "Task" [ 1188.962874] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.972115] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160620, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.990895] env[61986]: DEBUG nova.network.neutron [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating instance_info_cache with network_info: [{"id": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "address": "fa:16:3e:27:92:1e", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c570cf8-31", "ovs_interfaceid": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.298973] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Didn't find any instances for network info cache update. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1189.299253] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.299457] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.299676] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.299785] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.343691] env[61986]: DEBUG nova.objects.instance [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'numa_topology' on Instance uuid 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.358962] env[61986]: DEBUG oslo_vmware.api [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160619, 'name': PowerOffVM_Task, 'duration_secs': 0.283435} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.359339] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1189.359569] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1189.359880] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-daeda5b8-499d-4170-bd37-cc15462bc1b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.474152] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160620, 'name': Rename_Task, 'duration_secs': 0.148309} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.474445] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1189.474691] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a030552-e7dc-4aca-9564-8e2e43822ca3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.481764] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1189.481764] env[61986]: value = "task-1160622" [ 1189.481764] env[61986]: _type = "Task" [ 1189.481764] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.490115] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.493680] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.494007] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Instance network_info: |[{"id": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "address": "fa:16:3e:27:92:1e", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c570cf8-31", "ovs_interfaceid": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1189.494404] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:92:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c570cf8-315d-46b2-a056-e00ad031a6ab', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1189.502521] env[61986]: DEBUG oslo.service.loopingcall [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1189.502775] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1189.503013] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2359ae5-b0e7-4343-9b86-3082f48049cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.523704] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1189.523704] env[61986]: value = "task-1160623" [ 1189.523704] env[61986]: _type = "Task" [ 1189.523704] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.532113] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160623, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.546523] env[61986]: DEBUG nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Received event network-vif-plugged-0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1189.546760] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.546949] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.547157] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.547319] env[61986]: DEBUG nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] No waiting events found dispatching network-vif-plugged-0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1189.547464] env[61986]: WARNING nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Received unexpected event network-vif-plugged-0c570cf8-315d-46b2-a056-e00ad031a6ab for instance with vm_state building and task_state spawning. [ 1189.547616] env[61986]: DEBUG nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1189.547778] env[61986]: DEBUG nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing instance network info cache due to event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1189.547959] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.548113] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.548276] env[61986]: DEBUG nova.network.neutron [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1189.851521] env[61986]: DEBUG nova.objects.base [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Object Instance<4ba294fa-5a23-45eb-95ed-ac41d541ebf7> lazy-loaded attributes: resources,numa_topology {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1189.859159] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1189.859523] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1189.859739] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore1] d818b2ef-5c36-4df8-a3e0-1d48e027dd9d {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1189.860148] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8f41e0e-92b8-43cd-9164-1414010c26b9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.867047] env[61986]: DEBUG oslo_vmware.api [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1189.867047] env[61986]: value = "task-1160624" [ 1189.867047] env[61986]: _type = "Task" [ 1189.867047] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.876962] env[61986]: DEBUG oslo_vmware.api [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160624, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.939920] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464a0bd3-a6bb-4455-97ed-f22759f4eaaf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.948155] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844cab26-936a-4705-992d-0c2863eed191 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.979358] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da799a2f-4de7-468b-b542-6858c2249e31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.988207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5cc11f-311c-4da5-b1a0-367f24f182da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.996066] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160622, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.005604] env[61986]: DEBUG nova.compute.provider_tree [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.033904] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160623, 'name': CreateVM_Task} progress is 25%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.297233] env[61986]: DEBUG nova.network.neutron [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updated VIF entry in instance network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1190.297611] env[61986]: DEBUG nova.network.neutron [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0b5285ba-76", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.376857] env[61986]: DEBUG oslo_vmware.api [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160624, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148769} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.377174] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1190.377461] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1190.377681] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1190.377872] env[61986]: INFO nova.compute.manager [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Took 1.55 seconds to destroy the instance on the hypervisor. [ 1190.378327] env[61986]: DEBUG oslo.service.loopingcall [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1190.378327] env[61986]: DEBUG nova.compute.manager [-] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1190.378437] env[61986]: DEBUG nova.network.neutron [-] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1190.493924] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160622, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.511305] env[61986]: DEBUG nova.scheduler.client.report [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.539027] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160623, 'name': CreateVM_Task, 'duration_secs': 0.813081} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.539027] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1190.539027] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.539027] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.539027] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1190.539027] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bec9d6e-10ac-4aa2-b383-f627fb3243d8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.546027] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1190.546027] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ea8343-6f69-1aa7-28e3-00f2d8a06d68" [ 1190.546027] env[61986]: _type = "Task" [ 1190.546027] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.554580] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ea8343-6f69-1aa7-28e3-00f2d8a06d68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.800564] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.801094] env[61986]: DEBUG nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Received event network-changed-0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1190.801184] env[61986]: DEBUG nova.compute.manager [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Refreshing instance network info cache due to event network-changed-0c570cf8-315d-46b2-a056-e00ad031a6ab. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1190.801343] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Acquiring lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.801511] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Acquired lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.801679] env[61986]: DEBUG nova.network.neutron [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Refreshing network info cache for port 0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1190.993543] env[61986]: DEBUG oslo_vmware.api [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160622, 'name': PowerOnVM_Task, 'duration_secs': 1.063924} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.993880] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1190.994054] env[61986]: INFO nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Took 7.91 seconds to spawn the instance on the hypervisor. [ 1190.994244] env[61986]: DEBUG nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1190.994986] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a037f9-414a-464b-8e72-ed79bb0eec0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.016045] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.178s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.057304] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52ea8343-6f69-1aa7-28e3-00f2d8a06d68, 'name': SearchDatastore_Task, 'duration_secs': 0.011877} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.057605] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.057847] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1191.058091] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.058249] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.058434] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1191.058683] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc8a7e4d-31cc-46f6-8933-78c88fc37618 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.071654] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1191.071834] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1191.072551] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7f6d83b-eab3-4922-bfb5-815d41caf1ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.077581] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1191.077581] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52adcfbe-846c-4153-3858-81ea9b3169ae" [ 1191.077581] env[61986]: _type = "Task" [ 1191.077581] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.084959] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52adcfbe-846c-4153-3858-81ea9b3169ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.174530] env[61986]: DEBUG nova.network.neutron [-] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.511178] env[61986]: DEBUG nova.network.neutron [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updated VIF entry in instance network info cache for port 0c570cf8-315d-46b2-a056-e00ad031a6ab. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1191.511746] env[61986]: DEBUG nova.network.neutron [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating instance_info_cache with network_info: [{"id": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "address": "fa:16:3e:27:92:1e", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c570cf8-31", "ovs_interfaceid": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.513105] env[61986]: INFO nova.compute.manager [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Took 14.39 seconds to build instance. [ 1191.523035] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e12bb920-ced3-49a8-ba0c-4289e430963e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.821s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.588806] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52adcfbe-846c-4153-3858-81ea9b3169ae, 'name': SearchDatastore_Task, 'duration_secs': 0.036865} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.589672] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-712c5b00-e069-4ccd-88c7-463b00beefd9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.595668] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1191.595668] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5222e08b-bf01-d7ab-e0e9-77dbbd8d943e" [ 1191.595668] env[61986]: _type = "Task" [ 1191.595668] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.603324] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5222e08b-bf01-d7ab-e0e9-77dbbd8d943e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.628545] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.628545] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.628545] env[61986]: INFO nova.compute.manager [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Unshelving [ 1191.639621] env[61986]: DEBUG nova.compute.manager [req-4d4001b8-1040-49e2-8e67-16332d56ca36 req-556748f7-6b53-4bba-986e-aeea5f3fceb7 service nova] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Received event network-vif-deleted-938a568f-0336-4dc6-80f3-2f5994ef2ad2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1191.678232] env[61986]: INFO nova.compute.manager [-] [instance: d818b2ef-5c36-4df8-a3e0-1d48e027dd9d] Took 1.30 seconds to deallocate network for instance. [ 1192.015652] env[61986]: DEBUG oslo_concurrency.lockutils [req-e6a208f8-14f0-43d9-854d-85ba9b8867f2 req-e33a8e4b-8808-479f-b6ee-efd95f5f3427 service nova] Releasing lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.016208] env[61986]: DEBUG oslo_concurrency.lockutils [None req-d6c0917b-99f2-4588-accd-ee5b00f73f15 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.906s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.106854] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5222e08b-bf01-d7ab-e0e9-77dbbd8d943e, 'name': SearchDatastore_Task, 'duration_secs': 0.019607} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.107160] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.107420] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1192.107688] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91bd546d-0495-4e72-9bd3-149c07205d00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.114922] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1192.114922] env[61986]: value = "task-1160625" [ 1192.114922] env[61986]: _type = "Task" [ 1192.114922] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.122588] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.185284] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.185715] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.185996] env[61986]: DEBUG nova.objects.instance [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid d818b2ef-5c36-4df8-a3e0-1d48e027dd9d {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.191912] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.192162] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.192346] env[61986]: INFO nova.compute.manager [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Shelving [ 1192.626539] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498089} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.626717] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1192.627859] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1192.627859] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25f45b1e-ceb6-42e7-9c24-519a9c65a571 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.635496] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1192.635496] env[61986]: value = "task-1160626" [ 1192.635496] env[61986]: _type = "Task" [ 1192.635496] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.649275] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.657669] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.699629] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1192.699904] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93cb947d-a88e-41ed-93aa-9ecf32d33dd5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.708830] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1192.708830] env[61986]: value = "task-1160627" [ 1192.708830] env[61986]: _type = "Task" [ 1192.708830] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.717165] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.792322] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0354769d-052c-4f0a-867d-e973c3318d28 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.800194] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8d1497-8463-431e-8a45-d5d6b07cbde0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.831046] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745a9309-fb5a-433e-88ba-7085c84146c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.839379] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afb0a45-3dfb-41b8-8c7f-6ee8c5239bab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.854392] env[61986]: DEBUG nova.compute.provider_tree [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.858323] env[61986]: DEBUG nova.compute.manager [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Stashing vm_state: active {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1193.146778] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075806} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.147179] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1193.147830] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fac41b-0b88-4c2c-ba42-4458088467bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.169963] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1193.170267] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b2ffaf0-41bf-48e5-9e9f-0e1525322c97 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.191702] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1193.191702] env[61986]: value = "task-1160628" [ 1193.191702] env[61986]: _type = "Task" [ 1193.191702] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.201635] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160628, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.219105] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160627, 'name': PowerOffVM_Task, 'duration_secs': 0.219867} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.219426] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1193.220254] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8275e37-91a6-49c4-bc2a-2153bd299f73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.239287] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84ccfcd-8c38-4535-8226-14ac0d05d197 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.359524] env[61986]: DEBUG nova.scheduler.client.report [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.377743] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.702922] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160628, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.750400] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Creating Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1193.750733] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-16635797-d9e1-4284-a24d-146d95873e98 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.759840] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1193.759840] env[61986]: value = "task-1160629" [ 1193.759840] env[61986]: _type = "Task" [ 1193.759840] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.769563] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160629, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.868919] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.872026] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.214s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.872295] env[61986]: DEBUG nova.objects.instance [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'pci_requests' on Instance uuid 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.888158] env[61986]: INFO nova.scheduler.client.report [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance d818b2ef-5c36-4df8-a3e0-1d48e027dd9d [ 1194.208369] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160628, 'name': ReconfigVM_Task, 'duration_secs': 0.770286} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.208369] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1194.208369] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09e6b132-10c4-46a0-8d36-67f7685f1e93 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.215153] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1194.215153] env[61986]: value = "task-1160630" [ 1194.215153] env[61986]: _type = "Task" [ 1194.215153] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.223886] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160630, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.269580] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160629, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.376058] env[61986]: DEBUG nova.objects.instance [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'numa_topology' on Instance uuid 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.396640] env[61986]: DEBUG oslo_concurrency.lockutils [None req-aa8691c9-5ab4-4c33-9ac5-cce2ee080384 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "d818b2ef-5c36-4df8-a3e0-1d48e027dd9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.577s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.724642] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160630, 'name': Rename_Task, 'duration_secs': 0.170958} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.724936] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1194.725178] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-934cb474-baff-4833-8671-fc56b6a001c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.731390] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1194.731390] env[61986]: value = "task-1160631" [ 1194.731390] env[61986]: _type = "Task" [ 1194.731390] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.738590] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.769790] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160629, 'name': CreateSnapshot_Task, 'duration_secs': 0.898821} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.770089] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Created Snapshot of the VM instance {{(pid=61986) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1194.770743] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb1c169-d0e9-4bbb-9661-1396251113d5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.878093] env[61986]: INFO nova.compute.claims [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1195.085870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.086121] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.243096] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160631, 'name': PowerOnVM_Task} progress is 76%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.288661] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Creating linked-clone VM from snapshot {{(pid=61986) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1195.288661] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4c85bf79-5898-433c-9943-9f0b15db4d5a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.297929] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1195.297929] env[61986]: value = "task-1160632" [ 1195.297929] env[61986]: _type = "Task" [ 1195.297929] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.306943] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160632, 'name': CloneVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.589199] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1195.744586] env[61986]: DEBUG oslo_vmware.api [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160631, 'name': PowerOnVM_Task, 'duration_secs': 0.914762} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.744976] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1195.745286] env[61986]: INFO nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Took 8.53 seconds to spawn the instance on the hypervisor. [ 1195.745586] env[61986]: DEBUG nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1195.746780] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d2363c-a809-4ac5-9f77-bb57625819cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.810203] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160632, 'name': CloneVM_Task} progress is 94%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.995846] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6070b1a1-349e-487a-a83b-b9891cb37527 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.004875] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f33ead-cfee-47fc-993e-ae2d63e9ff9a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.035180] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca6d428-459a-467b-93f9-f5c88cfa97c8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.043187] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6dedd7-1edc-4af1-938a-590dff7ceb2a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.056599] env[61986]: DEBUG nova.compute.provider_tree [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.109181] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.268897] env[61986]: INFO nova.compute.manager [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Took 13.30 seconds to build instance. [ 1196.312423] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160632, 'name': CloneVM_Task} progress is 95%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.563076] env[61986]: DEBUG nova.scheduler.client.report [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1196.771661] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_power_states {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.771661] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8706cfda-ae1b-4e0a-8e16-9d64c03e7dd5 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.815s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.810944] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160632, 'name': CloneVM_Task, 'duration_secs': 1.176079} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.811812] env[61986]: INFO nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Created linked-clone VM from snapshot [ 1196.812264] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aec36d2-9276-402d-bd2a-5e6514d41913 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.820045] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Uploading image 3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1196.845447] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1196.845447] env[61986]: value = "vm-252495" [ 1196.845447] env[61986]: _type = "VirtualMachine" [ 1196.845447] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1196.845765] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-99ef8a61-0bb2-4975-bd1d-8c41d7e4d71f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.854306] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease: (returnval){ [ 1196.854306] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252f162-1fbe-f748-b54d-739bb2a8cd74" [ 1196.854306] env[61986]: _type = "HttpNfcLease" [ 1196.854306] env[61986]: } obtained for exporting VM: (result){ [ 1196.854306] env[61986]: value = "vm-252495" [ 1196.854306] env[61986]: _type = "VirtualMachine" [ 1196.854306] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1196.854764] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the lease: (returnval){ [ 1196.854764] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252f162-1fbe-f748-b54d-739bb2a8cd74" [ 1196.854764] env[61986]: _type = "HttpNfcLease" [ 1196.854764] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1196.864743] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1196.864743] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252f162-1fbe-f748-b54d-739bb2a8cd74" [ 1196.864743] env[61986]: _type = "HttpNfcLease" [ 1196.864743] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1196.898504] env[61986]: DEBUG nova.compute.manager [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Received event network-changed-0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1196.898668] env[61986]: DEBUG nova.compute.manager [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Refreshing instance network info cache due to event network-changed-0c570cf8-315d-46b2-a056-e00ad031a6ab. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1196.898886] env[61986]: DEBUG oslo_concurrency.lockutils [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] Acquiring lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.899043] env[61986]: DEBUG oslo_concurrency.lockutils [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] Acquired lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.899210] env[61986]: DEBUG nova.network.neutron [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Refreshing network info cache for port 0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1197.065759] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.194s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.068025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.690s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.120945] env[61986]: INFO nova.network.neutron [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating port 0b5285ba-76dd-4e56-aa09-5ac60d84235f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1197.275941] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Getting list of instances from cluster (obj){ [ 1197.275941] env[61986]: value = "domain-c8" [ 1197.275941] env[61986]: _type = "ClusterComputeResource" [ 1197.275941] env[61986]: } {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1197.275941] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b831d6-e221-4658-9080-2466be6bbe7f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.289897] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Got total of 5 instances {{(pid=61986) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1197.290066] env[61986]: WARNING nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] While synchronizing instance power states, found 6 instances in the database and 5 instances on the hypervisor. [ 1197.290212] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1197.290402] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid a6094cc4-7d17-4858-abae-972425241c8f {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1197.290557] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1197.290710] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1197.290863] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid cff8354d-1eb9-446c-8c63-6697264c373f {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1197.291016] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Triggering sync for uuid 66267549-1945-4c63-8b3b-c090a4ea1a28 {{(pid=61986) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1197.291398] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.291661] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.291961] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "a6094cc4-7d17-4858-abae-972425241c8f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.292168] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "a6094cc4-7d17-4858-abae-972425241c8f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.292413] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.292625] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.292825] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.293009] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.293187] env[61986]: INFO nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1197.293351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.293542] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.293721] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.294603] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55179845-659d-4181-8883-07403bcb1cb5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.297535] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3df115-e960-4929-a7c9-8d178975c9f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.300349] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547fc3df-dfe9-4a06-8da2-bd8da06e53e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.363958] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1197.363958] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252f162-1fbe-f748-b54d-739bb2a8cd74" [ 1197.363958] env[61986]: _type = "HttpNfcLease" [ 1197.363958] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1197.364300] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1197.364300] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5252f162-1fbe-f748-b54d-739bb2a8cd74" [ 1197.364300] env[61986]: _type = "HttpNfcLease" [ 1197.364300] env[61986]: }. {{(pid=61986) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1197.365060] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3859b8-631d-4c3d-920d-37892365803a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.373353] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526e94c1-932a-d56f-f3c2-addb016bef2a/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1197.373547] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526e94c1-932a-d56f-f3c2-addb016bef2a/disk-0.vmdk for reading. {{(pid=61986) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1197.468193] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4d512ee8-f991-4aa9-a985-723d904ba587 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.572875] env[61986]: INFO nova.compute.claims [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1197.653262] env[61986]: DEBUG nova.network.neutron [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updated VIF entry in instance network info cache for port 0c570cf8-315d-46b2-a056-e00ad031a6ab. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1197.653642] env[61986]: DEBUG nova.network.neutron [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating instance_info_cache with network_info: [{"id": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "address": "fa:16:3e:27:92:1e", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c570cf8-31", "ovs_interfaceid": "0c570cf8-315d-46b2-a056-e00ad031a6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.813640] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.814172] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.814540] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "a6094cc4-7d17-4858-abae-972425241c8f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.079337] env[61986]: INFO nova.compute.resource_tracker [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating resource usage from migration e913fad5-a9ea-4133-8fb9-e06a6e08c666 [ 1198.157853] env[61986]: DEBUG oslo_concurrency.lockutils [req-a14454e0-e4f5-4486-974b-fe7cedae905a req-608bf241-0d5c-4444-940b-a4cb3e08f395 service nova] Releasing lock "refresh_cache-66267549-1945-4c63-8b3b-c090a4ea1a28" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.201880] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8aa239-f1d7-4f82-8212-7dead10f117c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.210525] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9e12d7-8920-4c7c-8728-2ff150674970 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.243506] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02bdeb99-d427-4a16-8de2-dfb81f7d4b71 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.252337] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e589fd3-a946-49b1-bfc4-4c6ba5242648 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.267366] env[61986]: DEBUG nova.compute.provider_tree [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.538528] env[61986]: DEBUG nova.compute.manager [req-f8836153-95ca-4f9c-ac7d-fe0629c68428 req-f146fe2e-801d-4a79-b37b-23207315a3ca service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-vif-plugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1198.539254] env[61986]: DEBUG oslo_concurrency.lockutils [req-f8836153-95ca-4f9c-ac7d-fe0629c68428 req-f146fe2e-801d-4a79-b37b-23207315a3ca service nova] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.539673] env[61986]: DEBUG oslo_concurrency.lockutils [req-f8836153-95ca-4f9c-ac7d-fe0629c68428 req-f146fe2e-801d-4a79-b37b-23207315a3ca service nova] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.539841] env[61986]: DEBUG oslo_concurrency.lockutils [req-f8836153-95ca-4f9c-ac7d-fe0629c68428 req-f146fe2e-801d-4a79-b37b-23207315a3ca service nova] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.540049] env[61986]: DEBUG nova.compute.manager [req-f8836153-95ca-4f9c-ac7d-fe0629c68428 req-f146fe2e-801d-4a79-b37b-23207315a3ca service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] No waiting events found dispatching network-vif-plugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1198.540353] env[61986]: WARNING nova.compute.manager [req-f8836153-95ca-4f9c-ac7d-fe0629c68428 req-f146fe2e-801d-4a79-b37b-23207315a3ca service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received unexpected event network-vif-plugged-0b5285ba-76dd-4e56-aa09-5ac60d84235f for instance with vm_state shelved_offloaded and task_state spawning. [ 1198.643971] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.644686] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.644769] env[61986]: DEBUG nova.network.neutron [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1198.770604] env[61986]: DEBUG nova.scheduler.client.report [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1199.275441] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.207s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.275716] env[61986]: INFO nova.compute.manager [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Migrating [ 1199.282895] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.174s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.284435] env[61986]: INFO nova.compute.claims [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1199.396690] env[61986]: DEBUG nova.network.neutron [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.797194] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.797548] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.797548] env[61986]: DEBUG nova.network.neutron [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1199.899340] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.927919] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7c0d91e56fda5269a16857111b0a303a',container_format='bare',created_at=2024-10-07T06:48:46Z,direct_url=,disk_format='vmdk',id=b72c8918-6428-4315-9eb6-2e52bebfc404,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1318627242-shelved',owner='691d4935610e45e6a0bae5ca5f1197f3',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2024-10-07T06:48:59Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1199.928245] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1199.928413] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1199.928707] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1199.928879] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1199.929067] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1199.929306] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1199.929498] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1199.929691] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1199.929867] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1199.930080] env[61986]: DEBUG nova.virt.hardware [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1199.931020] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25812a63-5ecf-4bbc-94bb-b2a160f43eb3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.939927] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6920eb64-d01a-4feb-8b33-3064aeac62e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.955272] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:0e:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1768af3d-3317-4ef5-b484-0c2707d63de7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b5285ba-76dd-4e56-aa09-5ac60d84235f', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1199.967962] env[61986]: DEBUG oslo.service.loopingcall [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1199.968407] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1199.968777] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6c40f93-9473-40b8-a3dc-04b742348906 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.001825] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1200.001825] env[61986]: value = "task-1160634" [ 1200.001825] env[61986]: _type = "Task" [ 1200.001825] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.013135] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160634, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.421850] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95834f92-2178-4d73-9b87-8326673fb372 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.429768] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818705e7-8258-4913-9a81-76bd587f6638 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.464428] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cd5dd6-3aba-4528-b0d5-fb5b6b313338 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.472945] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36552bde-ccfb-4dd4-b65b-7b1a6ba261ba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.487857] env[61986]: DEBUG nova.compute.provider_tree [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.512167] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160634, 'name': CreateVM_Task, 'duration_secs': 0.375162} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.512354] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1200.513062] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.513243] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.513650] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1200.513903] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f2138f3-04f6-43a5-bc65-e138d9b7dab4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.519170] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1200.519170] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521da9b2-77fb-4290-e70d-26b9503f8b7d" [ 1200.519170] env[61986]: _type = "Task" [ 1200.519170] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.527647] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521da9b2-77fb-4290-e70d-26b9503f8b7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.538748] env[61986]: DEBUG nova.network.neutron [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [{"id": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "address": "fa:16:3e:e5:1f:66", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f357ac4-99", "ovs_interfaceid": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.569900] env[61986]: DEBUG nova.compute.manager [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1200.570126] env[61986]: DEBUG nova.compute.manager [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing instance network info cache due to event network-changed-0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1200.570344] env[61986]: DEBUG oslo_concurrency.lockutils [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] Acquiring lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.570489] env[61986]: DEBUG oslo_concurrency.lockutils [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] Acquired lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.570651] env[61986]: DEBUG nova.network.neutron [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Refreshing network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1200.991793] env[61986]: DEBUG nova.scheduler.client.report [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.031275] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.031611] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Processing image b72c8918-6428-4315-9eb6-2e52bebfc404 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1201.031870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.032041] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.032230] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1201.032496] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a201d7eb-00e1-4557-85df-80c130f2372c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.041617] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.052501] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1201.052753] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1201.053544] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62f0b0f9-9ee0-4719-bead-e3b85da6139d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.060261] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1201.060261] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5293ac83-a557-04d7-f387-c0c2f5db4bfe" [ 1201.060261] env[61986]: _type = "Task" [ 1201.060261] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.069242] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5293ac83-a557-04d7-f387-c0c2f5db4bfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.282945] env[61986]: DEBUG nova.network.neutron [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updated VIF entry in instance network info cache for port 0b5285ba-76dd-4e56-aa09-5ac60d84235f. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1201.283371] env[61986]: DEBUG nova.network.neutron [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [{"id": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "address": "fa:16:3e:0a:0e:fc", "network": {"id": "50fd50ab-6487-4eda-8384-dcc46afc4303", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-120605841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "691d4935610e45e6a0bae5ca5f1197f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1768af3d-3317-4ef5-b484-0c2707d63de7", "external-id": "nsx-vlan-transportzone-706", "segmentation_id": 706, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b5285ba-76", "ovs_interfaceid": "0b5285ba-76dd-4e56-aa09-5ac60d84235f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.496935] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.497986] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1201.571176] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1201.571438] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Fetch image to [datastore1] OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b/OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1201.571658] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Downloading stream optimized image b72c8918-6428-4315-9eb6-2e52bebfc404 to [datastore1] OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b/OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b.vmdk on the data store datastore1 as vApp {{(pid=61986) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1201.571842] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Downloading image file data b72c8918-6428-4315-9eb6-2e52bebfc404 to the ESX as VM named 'OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b' {{(pid=61986) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1201.641221] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1201.641221] env[61986]: value = "resgroup-9" [ 1201.641221] env[61986]: _type = "ResourcePool" [ 1201.641221] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1201.641579] env[61986]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-95e1ba4b-e3d6-4bea-bd73-1c256618cc48 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.662488] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease: (returnval){ [ 1201.662488] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52633b3d-a90e-964c-e741-6ec4e0c85fdd" [ 1201.662488] env[61986]: _type = "HttpNfcLease" [ 1201.662488] env[61986]: } obtained for vApp import into resource pool (val){ [ 1201.662488] env[61986]: value = "resgroup-9" [ 1201.662488] env[61986]: _type = "ResourcePool" [ 1201.662488] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1201.662823] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the lease: (returnval){ [ 1201.662823] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52633b3d-a90e-964c-e741-6ec4e0c85fdd" [ 1201.662823] env[61986]: _type = "HttpNfcLease" [ 1201.662823] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1201.670186] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1201.670186] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52633b3d-a90e-964c-e741-6ec4e0c85fdd" [ 1201.670186] env[61986]: _type = "HttpNfcLease" [ 1201.670186] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1201.786798] env[61986]: DEBUG oslo_concurrency.lockutils [req-b1426019-c2e9-4436-9b13-244287048f0f req-a69b0d71-8558-45c3-b702-9729aa1c1f47 service nova] Releasing lock "refresh_cache-4ba294fa-5a23-45eb-95ed-ac41d541ebf7" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.003738] env[61986]: DEBUG nova.compute.utils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1202.005581] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1202.005709] env[61986]: DEBUG nova.network.neutron [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1202.046512] env[61986]: DEBUG nova.policy [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '207c795b4a1241e68f9873bdb5cafeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d92ccda87d241068595992a1b8d3029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1202.172031] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1202.172031] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52633b3d-a90e-964c-e741-6ec4e0c85fdd" [ 1202.172031] env[61986]: _type = "HttpNfcLease" [ 1202.172031] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1202.307194] env[61986]: DEBUG nova.network.neutron [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Successfully created port: 376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1202.508773] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1202.558401] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33456112-c9bf-479a-8d14-43053159ae49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.580146] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1202.671873] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1202.671873] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52633b3d-a90e-964c-e741-6ec4e0c85fdd" [ 1202.671873] env[61986]: _type = "HttpNfcLease" [ 1202.671873] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1202.672191] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1202.672191] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52633b3d-a90e-964c-e741-6ec4e0c85fdd" [ 1202.672191] env[61986]: _type = "HttpNfcLease" [ 1202.672191] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1202.672943] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b22b3f-4ba5-47ed-a0d0-0566e10bb4cd {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.680125] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac264-7965-8fbe-77b8-da6d4a4cf1a3/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1202.680323] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac264-7965-8fbe-77b8-da6d4a4cf1a3/disk-0.vmdk. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1202.743417] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f41fcbd2-0dbf-4436-85c5-98bc5f8b24b1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.091564] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1203.091971] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0342d782-7249-486d-980a-884790e2738e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.098774] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1203.098774] env[61986]: value = "task-1160636" [ 1203.098774] env[61986]: _type = "Task" [ 1203.098774] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.106927] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.518059] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1203.549144] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1203.549447] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1203.549718] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1203.549962] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1203.550079] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1203.550232] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1203.550441] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1203.550609] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1203.550805] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1203.550977] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1203.551186] env[61986]: DEBUG nova.virt.hardware [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1203.552307] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6835fbad-75d0-49f8-9700-9bc385bd0ff8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.564278] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e503e6a-917a-47c9-9c61-f40748c0118e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.608191] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160636, 'name': PowerOffVM_Task, 'duration_secs': 0.317746} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.608527] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1203.608744] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1203.841066] env[61986]: DEBUG nova.compute.manager [req-e893c312-1e09-43df-9add-0fe0feeccc04 req-99e8db1c-007a-40f5-92f9-a4ec425dd98d service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Received event network-vif-plugged-376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1203.841066] env[61986]: DEBUG oslo_concurrency.lockutils [req-e893c312-1e09-43df-9add-0fe0feeccc04 req-99e8db1c-007a-40f5-92f9-a4ec425dd98d service nova] Acquiring lock "3fb03cbe-3c13-4fce-8556-aaec6030b264-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.841066] env[61986]: DEBUG oslo_concurrency.lockutils [req-e893c312-1e09-43df-9add-0fe0feeccc04 req-99e8db1c-007a-40f5-92f9-a4ec425dd98d service nova] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.841066] env[61986]: DEBUG oslo_concurrency.lockutils [req-e893c312-1e09-43df-9add-0fe0feeccc04 req-99e8db1c-007a-40f5-92f9-a4ec425dd98d service nova] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.841066] env[61986]: DEBUG nova.compute.manager [req-e893c312-1e09-43df-9add-0fe0feeccc04 req-99e8db1c-007a-40f5-92f9-a4ec425dd98d service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] No waiting events found dispatching network-vif-plugged-376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1203.841066] env[61986]: WARNING nova.compute.manager [req-e893c312-1e09-43df-9add-0fe0feeccc04 req-99e8db1c-007a-40f5-92f9-a4ec425dd98d service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Received unexpected event network-vif-plugged-376c7f76-71f0-4f6f-8514-0769daaecb5b for instance with vm_state building and task_state spawning. [ 1204.036211] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1204.036520] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac264-7965-8fbe-77b8-da6d4a4cf1a3/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1204.037627] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0722b85-7797-4003-94df-101c270ac3c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.044281] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac264-7965-8fbe-77b8-da6d4a4cf1a3/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1204.044454] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac264-7965-8fbe-77b8-da6d4a4cf1a3/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1204.044783] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-134aea7b-a555-4a55-a209-f8decf9518d2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.115249] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1204.115652] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1204.115652] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1204.115849] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1204.115999] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1204.116201] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1204.116415] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1204.116643] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1204.116864] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1204.117060] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1204.117244] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1204.122535] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4bde027-e680-4c42-8220-dc4ab25e02d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.139957] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1204.139957] env[61986]: value = "task-1160637" [ 1204.139957] env[61986]: _type = "Task" [ 1204.139957] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.148805] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160637, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.382538] env[61986]: DEBUG oslo_vmware.rw_handles [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac264-7965-8fbe-77b8-da6d4a4cf1a3/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1204.382930] env[61986]: INFO nova.virt.vmwareapi.images [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Downloaded image file data b72c8918-6428-4315-9eb6-2e52bebfc404 [ 1204.387398] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1068bf-7253-4615-9261-5cb31f166a34 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.401432] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d524a46-f196-4afa-a05c-bf8728ffea43 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.452032] env[61986]: INFO nova.virt.vmwareapi.images [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] The imported VM was unregistered [ 1204.452032] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1204.452032] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Creating directory with path [datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1204.452032] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7df4a134-e42c-4133-b5c2-a1f9424a875a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.456810] env[61986]: DEBUG nova.network.neutron [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Successfully updated port: 376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1204.467301] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Created directory with path [datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1204.467655] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b/OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b.vmdk to [datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk. {{(pid=61986) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1204.468101] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-1acc0984-291c-42a2-82a6-ffe62afb0da5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.476196] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1204.476196] env[61986]: value = "task-1160639" [ 1204.476196] env[61986]: _type = "Task" [ 1204.476196] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.486333] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160639, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.492934] env[61986]: DEBUG nova.compute.manager [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Received event network-changed-376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1204.493156] env[61986]: DEBUG nova.compute.manager [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Refreshing instance network info cache due to event network-changed-376c7f76-71f0-4f6f-8514-0769daaecb5b. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1204.493424] env[61986]: DEBUG oslo_concurrency.lockutils [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] Acquiring lock "refresh_cache-3fb03cbe-3c13-4fce-8556-aaec6030b264" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.493580] env[61986]: DEBUG oslo_concurrency.lockutils [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] Acquired lock "refresh_cache-3fb03cbe-3c13-4fce-8556-aaec6030b264" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.493676] env[61986]: DEBUG nova.network.neutron [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Refreshing network info cache for port 376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1204.649899] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160637, 'name': ReconfigVM_Task, 'duration_secs': 0.299644} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.650189] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1204.959627] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "refresh_cache-3fb03cbe-3c13-4fce-8556-aaec6030b264" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.990767] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160639, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.036158] env[61986]: DEBUG nova.network.neutron [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1205.120527] env[61986]: DEBUG nova.network.neutron [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.128253] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526e94c1-932a-d56f-f3c2-addb016bef2a/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1205.129306] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db153ea-ef77-4fbc-b684-34f770109c1b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.139591] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526e94c1-932a-d56f-f3c2-addb016bef2a/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1205.139802] env[61986]: ERROR oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526e94c1-932a-d56f-f3c2-addb016bef2a/disk-0.vmdk due to incomplete transfer. [ 1205.140060] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-95930c86-f224-4b62-93b8-b7f8a0995b1f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.150883] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526e94c1-932a-d56f-f3c2-addb016bef2a/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1205.151100] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Uploaded image 3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 to the Glance image server {{(pid=61986) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1205.153640] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Destroying the VM {{(pid=61986) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1205.153934] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2085f911-f07c-4445-9caf-d5c71c4973cc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.157469] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1205.157658] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1205.157825] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1205.158027] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1205.158188] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1205.158340] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1205.158551] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1205.158715] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1205.158886] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1205.159066] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1205.159252] env[61986]: DEBUG nova.virt.hardware [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1205.164689] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Reconfiguring VM instance instance-00000072 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1205.164993] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-786f58e0-6c91-4c1c-9e26-cb2afcb3eac0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.187087] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1205.187087] env[61986]: value = "task-1160640" [ 1205.187087] env[61986]: _type = "Task" [ 1205.187087] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.188202] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1205.188202] env[61986]: value = "task-1160641" [ 1205.188202] env[61986]: _type = "Task" [ 1205.188202] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.205878] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160641, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.206155] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160640, 'name': Destroy_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.489802] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160639, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.623560] env[61986]: DEBUG oslo_concurrency.lockutils [req-866bad25-86e6-4f8e-a0c5-d2e759220ad8 req-7af022a3-7aa2-4b39-b226-1095f54966e5 service nova] Releasing lock "refresh_cache-3fb03cbe-3c13-4fce-8556-aaec6030b264" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1205.624066] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "refresh_cache-3fb03cbe-3c13-4fce-8556-aaec6030b264" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.624248] env[61986]: DEBUG nova.network.neutron [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1205.703505] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160640, 'name': Destroy_Task, 'duration_secs': 0.383048} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.706653] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Destroyed the VM [ 1205.706925] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deleting Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1205.707240] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160641, 'name': ReconfigVM_Task, 'duration_secs': 0.204936} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.707454] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a2908519-722d-4ae0-a4db-7bb2d665adee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.709423] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Reconfigured VM instance instance-00000072 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1205.710219] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f623738-b29c-440f-94a7-b71f5c68b379 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.729662] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1205.729662] env[61986]: value = "task-1160642" [ 1205.729662] env[61986]: _type = "Task" [ 1205.729662] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.737645] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f/cff8354d-1eb9-446c-8c63-6697264c373f.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1205.738079] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97398a2f-dea3-4c94-88ec-716330151626 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.762626] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160642, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.764061] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1205.764061] env[61986]: value = "task-1160643" [ 1205.764061] env[61986]: _type = "Task" [ 1205.764061] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.773716] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160643, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.991600] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160639, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.156973] env[61986]: DEBUG nova.network.neutron [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1206.249309] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160642, 'name': RemoveSnapshot_Task, 'duration_secs': 0.386311} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.249601] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deleted Snapshot of the VM instance {{(pid=61986) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1206.249913] env[61986]: DEBUG nova.compute.manager [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1206.250701] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb8221a-3886-42e0-9b93-baa5236ab090 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.275040] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160643, 'name': ReconfigVM_Task, 'duration_secs': 0.391679} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.275361] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Reconfigured VM instance instance-00000072 to attach disk [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f/cff8354d-1eb9-446c-8c63-6697264c373f.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1206.275645] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 50 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1206.310228] env[61986]: DEBUG nova.network.neutron [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Updating instance_info_cache with network_info: [{"id": "376c7f76-71f0-4f6f-8514-0769daaecb5b", "address": "fa:16:3e:3f:1a:95", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap376c7f76-71", "ovs_interfaceid": "376c7f76-71f0-4f6f-8514-0769daaecb5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.490756] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160639, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.765146] env[61986]: INFO nova.compute.manager [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Shelve offloading [ 1206.767090] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1206.767425] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f334a6ed-ad8d-4c93-8094-10c299640934 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.777218] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1206.777218] env[61986]: value = "task-1160644" [ 1206.777218] env[61986]: _type = "Task" [ 1206.777218] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.782746] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c9a816-c133-4e2a-a7bb-b0b42bedf0b7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.807206] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1206.807426] env[61986]: DEBUG nova.compute.manager [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1206.808263] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104c6d87-3821-4299-b4df-bea746e7508e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.811405] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e3e601-ff13-465a-9fb7-28fb42b0f40e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.814355] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "refresh_cache-3fb03cbe-3c13-4fce-8556-aaec6030b264" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1206.814658] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Instance network_info: |[{"id": "376c7f76-71f0-4f6f-8514-0769daaecb5b", "address": "fa:16:3e:3f:1a:95", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap376c7f76-71", "ovs_interfaceid": "376c7f76-71f0-4f6f-8514-0769daaecb5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1206.815644] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:1a:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a7188cb6-c9d7-449a-9362-ddb61a31feaf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '376c7f76-71f0-4f6f-8514-0769daaecb5b', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1206.823247] env[61986]: DEBUG oslo.service.loopingcall [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1206.837259] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1206.837589] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 67 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1206.842799] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b6e4957-2203-492a-af3a-ec0d299a4a67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.857971] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1206.858170] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.858351] env[61986]: DEBUG nova.network.neutron [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1206.865329] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1206.865329] env[61986]: value = "task-1160645" [ 1206.865329] env[61986]: _type = "Task" [ 1206.865329] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.875751] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160645, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.987671] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160639, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.434064} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.987998] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b/OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b.vmdk to [datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk. [ 1206.988210] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Cleaning up location [datastore1] OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1206.988379] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a666c740-6813-4d73-bc6f-df5ddb58a72b {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1206.988628] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3d38cb5-96c8-41a0-952e-dc6e5a7f1aea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.994943] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1206.994943] env[61986]: value = "task-1160646" [ 1206.994943] env[61986]: _type = "Task" [ 1206.994943] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.002590] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.374871] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160645, 'name': CreateVM_Task, 'duration_secs': 0.344959} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.375201] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1207.375666] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.375841] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.376188] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1207.376435] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ffa1696-2126-496d-871a-1162e12689a2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.380643] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1207.380643] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529a095c-34ee-3387-4039-4a0d806e4789" [ 1207.380643] env[61986]: _type = "Task" [ 1207.380643] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.388259] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529a095c-34ee-3387-4039-4a0d806e4789, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.423367] env[61986]: DEBUG nova.network.neutron [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Port 9f357ac4-99e2-44ce-8f2c-3de7619c93fa binding to destination host cpu-1 is already ACTIVE {{(pid=61986) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1207.504837] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160646, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037316} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.505111] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1207.505285] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.505525] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk to [datastore1] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1207.505777] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44885297-8b2f-4e14-9413-aaaf3895dbdb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.512374] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1207.512374] env[61986]: value = "task-1160647" [ 1207.512374] env[61986]: _type = "Task" [ 1207.512374] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.519842] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160647, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.584030] env[61986]: DEBUG nova.network.neutron [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.892592] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]529a095c-34ee-3387-4039-4a0d806e4789, 'name': SearchDatastore_Task, 'duration_secs': 0.009238} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.893018] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.893183] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1207.893426] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.893576] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.893758] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1207.894089] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e8766a6-e86b-4816-8561-78b844e2ba70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.911247] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1207.911455] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1207.912411] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-057fbb2c-b399-4bd1-b221-1846a246ec93 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.920983] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1207.920983] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52108fef-83ab-1b99-d704-534bc0e42b56" [ 1207.920983] env[61986]: _type = "Task" [ 1207.920983] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.933370] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52108fef-83ab-1b99-d704-534bc0e42b56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.025561] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160647, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.087139] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.433540] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52108fef-83ab-1b99-d704-534bc0e42b56, 'name': SearchDatastore_Task, 'duration_secs': 0.081305} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.442296] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a804f63-1133-4515-9d85-47a4d491c47c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.453984] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1208.454496] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.454757] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.454945] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.457020] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e04d74-f0d2-47f7-bf3a-073e5a92e5ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.461751] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1208.461751] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b4668f-5fbb-a34c-72be-ce6f51496505" [ 1208.461751] env[61986]: _type = "Task" [ 1208.461751] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.468099] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1208.469573] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39dd8d6d-7298-4bc1-8bda-7f17a3e2d9b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.475884] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b4668f-5fbb-a34c-72be-ce6f51496505, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.524449] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160647, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.532940] env[61986]: DEBUG nova.compute.manager [req-02f43192-fab5-49ac-88e2-a8165880c418 req-17b979f3-adc6-4beb-8036-a6db2246923d service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-vif-unplugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1208.533337] env[61986]: DEBUG oslo_concurrency.lockutils [req-02f43192-fab5-49ac-88e2-a8165880c418 req-17b979f3-adc6-4beb-8036-a6db2246923d service nova] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.533645] env[61986]: DEBUG oslo_concurrency.lockutils [req-02f43192-fab5-49ac-88e2-a8165880c418 req-17b979f3-adc6-4beb-8036-a6db2246923d service nova] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.533809] env[61986]: DEBUG oslo_concurrency.lockutils [req-02f43192-fab5-49ac-88e2-a8165880c418 req-17b979f3-adc6-4beb-8036-a6db2246923d service nova] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.533995] env[61986]: DEBUG nova.compute.manager [req-02f43192-fab5-49ac-88e2-a8165880c418 req-17b979f3-adc6-4beb-8036-a6db2246923d service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] No waiting events found dispatching network-vif-unplugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1208.534186] env[61986]: WARNING nova.compute.manager [req-02f43192-fab5-49ac-88e2-a8165880c418 req-17b979f3-adc6-4beb-8036-a6db2246923d service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received unexpected event network-vif-unplugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 for instance with vm_state shelved and task_state shelving_offloading. [ 1208.540732] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1208.540940] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1208.541139] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleting the datastore file [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.541697] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36720e0e-ee7f-4ada-b0a8-7b91d27418b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.559878] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1208.559878] env[61986]: value = "task-1160649" [ 1208.559878] env[61986]: _type = "Task" [ 1208.559878] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.570316] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.975642] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52b4668f-5fbb-a34c-72be-ce6f51496505, 'name': SearchDatastore_Task, 'duration_secs': 0.056192} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.975642] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.975864] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3fb03cbe-3c13-4fce-8556-aaec6030b264/3fb03cbe-3c13-4fce-8556-aaec6030b264.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1208.976079] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f01e113-5e8b-468b-b2a8-9697a1f51c25 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.984928] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1208.984928] env[61986]: value = "task-1160650" [ 1208.984928] env[61986]: _type = "Task" [ 1208.984928] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.995159] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.025013] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160647, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.071536] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.214045] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1209.214224] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1209.499461] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.520799] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1209.520996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.521210] env[61986]: DEBUG nova.network.neutron [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1209.528108] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160647, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.573179] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.998955] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160650, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.028139] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160647, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.173427} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.028487] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b72c8918-6428-4315-9eb6-2e52bebfc404/b72c8918-6428-4315-9eb6-2e52bebfc404.vmdk to [datastore1] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1210.031747] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98669cbd-ed8e-473e-8b53-c4b35d20c2a0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.058023] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1210.059025] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c10efee-ba86-4a3a-92e1-3086722ef0c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.087094] env[61986]: DEBUG oslo_vmware.api [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.184796} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.088744] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1210.088994] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1210.089247] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1210.091942] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1210.091942] env[61986]: value = "task-1160651" [ 1210.091942] env[61986]: _type = "Task" [ 1210.091942] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.101144] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160651, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.122798] env[61986]: INFO nova.scheduler.client.report [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted allocations for instance c5a4fd56-25e2-4a95-8d2a-58361ef65e86 [ 1210.214598] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.214598] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.250527] env[61986]: DEBUG nova.network.neutron [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [{"id": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "address": "fa:16:3e:e5:1f:66", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f357ac4-99", "ovs_interfaceid": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.497412] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160650, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.088582} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.498041] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 3fb03cbe-3c13-4fce-8556-aaec6030b264/3fb03cbe-3c13-4fce-8556-aaec6030b264.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1210.498041] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1210.498231] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-963e7449-77fe-41e2-b674-96a315d80edb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.504652] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1210.504652] env[61986]: value = "task-1160652" [ 1210.504652] env[61986]: _type = "Task" [ 1210.504652] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.512499] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160652, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.561583] env[61986]: DEBUG nova.compute.manager [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1210.561806] env[61986]: DEBUG nova.compute.manager [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing instance network info cache due to event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1210.562123] env[61986]: DEBUG oslo_concurrency.lockutils [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.562289] env[61986]: DEBUG oslo_concurrency.lockutils [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.562482] env[61986]: DEBUG nova.network.neutron [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1210.602468] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160651, 'name': ReconfigVM_Task, 'duration_secs': 0.309935} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.602769] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7/4ba294fa-5a23-45eb-95ed-ac41d541ebf7.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1210.603658] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5c6ccd4-d9b8-4a88-9477-683e73345d4b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.609833] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1210.609833] env[61986]: value = "task-1160653" [ 1210.609833] env[61986]: _type = "Task" [ 1210.609833] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.617600] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160653, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.626703] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.626945] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.627180] env[61986]: DEBUG nova.objects.instance [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'resources' on Instance uuid c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1210.719906] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.753773] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.016308] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160652, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066491} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.020021] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1211.020021] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f67a198-204f-4e3e-8149-2ea3c8d2e759 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.046022] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 3fb03cbe-3c13-4fce-8556-aaec6030b264/3fb03cbe-3c13-4fce-8556-aaec6030b264.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1211.046353] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aace791f-2d0a-4c0c-a35d-8700eb63728d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.067849] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1211.067849] env[61986]: value = "task-1160654" [ 1211.067849] env[61986]: _type = "Task" [ 1211.067849] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.077266] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160654, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.119358] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160653, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.129522] env[61986]: DEBUG nova.objects.instance [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'numa_topology' on Instance uuid c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.274870] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08335065-ff2e-4858-ac7c-0e8e920aed05 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.278973] env[61986]: DEBUG nova.network.neutron [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updated VIF entry in instance network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1211.279311] env[61986]: DEBUG nova.network.neutron [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": null, "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap37976b99-2e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.295062] env[61986]: DEBUG oslo_concurrency.lockutils [req-060b16af-20ad-47d4-aaff-5dba003191e1 req-fc1c4d42-6d11-4dbc-9726-1d999de6a9bd service nova] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.295890] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde0f26b-907b-4c81-af3c-9c2186ec9d73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.303071] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 83 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1211.578782] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.619711] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160653, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.632651] env[61986]: DEBUG nova.objects.base [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1211.736638] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63dcd855-da0c-4dd6-af08-8e234f49720c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.751236] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4162149-0749-4a6c-94a9-27c28bb79d91 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.781378] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bca13e-fa5b-4bae-91d9-78a4d0d19e65 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.788386] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648fde0c-5a66-4f4b-b9d0-3bbc233e3687 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.801066] env[61986]: DEBUG nova.compute.provider_tree [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.809145] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1211.809862] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f64cc53-297c-4d22-b645-67ef94e324b4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.815781] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1211.815781] env[61986]: value = "task-1160655" [ 1211.815781] env[61986]: _type = "Task" [ 1211.815781] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.822954] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.053701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.079336] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160654, 'name': ReconfigVM_Task, 'duration_secs': 0.875647} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.079623] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 3fb03cbe-3c13-4fce-8556-aaec6030b264/3fb03cbe-3c13-4fce-8556-aaec6030b264.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1212.080282] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbd7b473-d0f5-42d6-9d60-df2003c87494 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.086999] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1212.086999] env[61986]: value = "task-1160656" [ 1212.086999] env[61986]: _type = "Task" [ 1212.086999] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.095090] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160656, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.119471] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160653, 'name': Rename_Task, 'duration_secs': 1.142016} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.119655] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1212.119883] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef91f680-237d-4acf-8c12-f726bd6c1c37 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.125964] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1212.125964] env[61986]: value = "task-1160657" [ 1212.125964] env[61986]: _type = "Task" [ 1212.125964] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.133080] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.304685] env[61986]: DEBUG nova.scheduler.client.report [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1212.328331] env[61986]: DEBUG oslo_vmware.api [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160655, 'name': PowerOnVM_Task, 'duration_secs': 0.424338} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.328655] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1212.328864] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a052a5fc-849e-4ca5-8e24-33c3c12941ae tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance 'cff8354d-1eb9-446c-8c63-6697264c373f' progress to 100 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1212.599558] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160656, 'name': Rename_Task, 'duration_secs': 0.156194} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.599914] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1212.600191] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff464af1-1ba4-4207-a6d9-471e3348b446 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.607808] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1212.607808] env[61986]: value = "task-1160658" [ 1212.607808] env[61986]: _type = "Task" [ 1212.607808] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.617362] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.636622] env[61986]: DEBUG oslo_vmware.api [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160657, 'name': PowerOnVM_Task, 'duration_secs': 0.421232} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.636965] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1212.745068] env[61986]: DEBUG nova.compute.manager [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1212.746100] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3addfe1c-3d0e-4b9f-92a0-23c93fd03f1b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.809424] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.182s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.811905] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.092s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.812169] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.812283] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1212.813377] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194e933e-643f-4ff2-b1e2-722493e42594 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.821399] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58aa68b9-56eb-451b-b1b2-b25adf52ad0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.838569] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab2cb21-8b13-479f-aebc-2cc295e909d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.845734] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfbdc91-331e-41d5-b599-0028a8a155b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.877021] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180658MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1212.877205] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.877379] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.118529] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160658, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.263081] env[61986]: DEBUG oslo_concurrency.lockutils [None req-2b80132f-b5d7-4ca6-b1b4-33dba0ed455e tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.635s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.264854] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.972s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.264854] env[61986]: INFO nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] During sync_power_state the instance has a pending task (spawning). Skip. [ 1213.264854] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.317330] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8bc90b0-5463-4351-a74d-b03207f723de tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.125s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.318212] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 16.025s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.318409] env[61986]: INFO nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1213.318586] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.318994] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.265s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.319186] env[61986]: INFO nova.compute.manager [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Unshelving [ 1213.619891] env[61986]: DEBUG oslo_vmware.api [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160658, 'name': PowerOnVM_Task, 'duration_secs': 0.690849} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.620354] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1213.620487] env[61986]: INFO nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Took 10.10 seconds to spawn the instance on the hypervisor. [ 1213.620699] env[61986]: DEBUG nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1213.621558] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a444a0-291c-4637-ae5f-36af3f287584 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.885633] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Applying migration context for instance cff8354d-1eb9-446c-8c63-6697264c373f as it has an incoming, in-progress migration e913fad5-a9ea-4133-8fb9-e06a6e08c666. Migration status is finished {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1213.886615] env[61986]: INFO nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating resource usage from migration e913fad5-a9ea-4133-8fb9-e06a6e08c666 [ 1213.906553] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1213.906705] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a6094cc4-7d17-4858-abae-972425241c8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1213.906829] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 66267549-1945-4c63-8b3b-c090a4ea1a28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1213.907180] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1213.907180] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Migration e913fad5-a9ea-4133-8fb9-e06a6e08c666 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1213.907301] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance cff8354d-1eb9-446c-8c63-6697264c373f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1213.907337] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 3fb03cbe-3c13-4fce-8556-aaec6030b264 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1214.138872] env[61986]: INFO nova.compute.manager [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Took 18.04 seconds to build instance. [ 1214.341416] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.410258] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance c5a4fd56-25e2-4a95-8d2a-58361ef65e86 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1214.411206] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1214.411206] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1214.520077] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fd2180-fc61-45c6-b711-2bfaec6586d3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.528186] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b3c7d0-0f71-41b7-b904-bff9e6acefd6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.559603] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.559926] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.560165] env[61986]: DEBUG nova.compute.manager [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Going to confirm migration 6 {{(pid=61986) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1214.563145] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d1ee68-db6c-43e8-b383-8cbfbe35e203 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.570915] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce39c43-7b32-436d-a310-e190b31b42af {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.584883] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.641254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-facee987-0de1-48af-849a-29c70f255293 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.555s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.088286] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1215.123921] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1215.124173] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquired lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.124358] env[61986]: DEBUG nova.network.neutron [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1215.124548] env[61986]: DEBUG nova.objects.instance [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'info_cache' on Instance uuid cff8354d-1eb9-446c-8c63-6697264c373f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.185861] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.186142] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.186333] env[61986]: DEBUG nova.compute.manager [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1215.187256] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970abe0a-2858-46ad-94ac-95d5a5888517 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.194652] env[61986]: DEBUG nova.compute.manager [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61986) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1215.195208] env[61986]: DEBUG nova.objects.instance [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'flavor' on Instance uuid 3fb03cbe-3c13-4fce-8556-aaec6030b264 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.592761] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1215.593023] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.716s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.593317] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.252s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.593540] env[61986]: DEBUG nova.objects.instance [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'pci_requests' on Instance uuid c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.700410] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1215.700611] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fdb9125-2696-4ad9-9ebd-9f0759042cb1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.707539] env[61986]: DEBUG oslo_vmware.api [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1215.707539] env[61986]: value = "task-1160659" [ 1215.707539] env[61986]: _type = "Task" [ 1215.707539] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.716704] env[61986]: DEBUG oslo_vmware.api [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160659, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.097449] env[61986]: DEBUG nova.objects.instance [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'numa_topology' on Instance uuid c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.217547] env[61986]: DEBUG oslo_vmware.api [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160659, 'name': PowerOffVM_Task, 'duration_secs': 0.240761} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.217825] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1216.218011] env[61986]: DEBUG nova.compute.manager [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1216.219350] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191f8002-b03b-4083-b961-dc751a8710c3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.324046] env[61986]: DEBUG nova.network.neutron [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [{"id": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "address": "fa:16:3e:e5:1f:66", "network": {"id": "d6b8e242-5fa8-4a3a-a357-b524e86c527b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127029727-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2e0f7d6b6be498eaec797aa64aaea7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f357ac4-99", "ovs_interfaceid": "9f357ac4-99e2-44ce-8f2c-3de7619c93fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.593789] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1216.594049] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1216.594201] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1216.594319] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Rebuilding the list of instances to heal {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1216.600430] env[61986]: INFO nova.compute.claims [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1216.730741] env[61986]: DEBUG oslo_concurrency.lockutils [None req-924102b4-061e-44b4-a043-63cc5900734f tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.826613] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Releasing lock "refresh_cache-cff8354d-1eb9-446c-8c63-6697264c373f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.826897] env[61986]: DEBUG nova.objects.instance [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lazy-loading 'migration_context' on Instance uuid cff8354d-1eb9-446c-8c63-6697264c373f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.099733] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Skipping network cache update for instance because it is being deleted. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1217.124315] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.124462] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.124611] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1217.124917] env[61986]: DEBUG nova.objects.instance [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lazy-loading 'info_cache' on Instance uuid 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.330983] env[61986]: DEBUG nova.objects.base [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61986) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1217.331698] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca9c130-e36d-4fe5-8090-ded0caed6d11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.351564] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bce53fa4-3155-440e-9a50-ac3f6381235a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.356946] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1217.356946] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52491213-4801-95aa-e83b-28138342ce4f" [ 1217.356946] env[61986]: _type = "Task" [ 1217.356946] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.364831] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52491213-4801-95aa-e83b-28138342ce4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.724233] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49abe5aa-3a81-4969-b4ce-aac3e588f9be {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.732939] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5f5f58-d518-4417-94cc-54f5e7455e6a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.763266] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73ce147-4795-4ac0-85fa-a4576173f7ac {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.771173] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461300ed-9a24-4d0d-bf05-27cef0d8d9cc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.786026] env[61986]: DEBUG nova.compute.provider_tree [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.791008] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.791264] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.791475] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "3fb03cbe-3c13-4fce-8556-aaec6030b264-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.791702] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.791885] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.794097] env[61986]: INFO nova.compute.manager [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Terminating instance [ 1217.796320] env[61986]: DEBUG nova.compute.manager [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1217.796520] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1217.797382] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e351ff3a-015d-423d-84a5-bc06971868ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.805408] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1217.805921] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ac34396-c80e-454c-b602-e9f7c799e70f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.867527] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52491213-4801-95aa-e83b-28138342ce4f, 'name': SearchDatastore_Task, 'duration_secs': 0.007776} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.869183] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.869491] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1217.869678] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1217.869862] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore1] 3fb03cbe-3c13-4fce-8556-aaec6030b264 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1217.870134] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-857e6b9a-1e0d-4331-99fc-3b279185217e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.876965] env[61986]: DEBUG oslo_vmware.api [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1217.876965] env[61986]: value = "task-1160661" [ 1217.876965] env[61986]: _type = "Task" [ 1217.876965] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.886193] env[61986]: DEBUG oslo_vmware.api [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160661, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.288946] env[61986]: DEBUG nova.scheduler.client.report [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1218.386755] env[61986]: DEBUG oslo_vmware.api [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160661, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135429} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.387024] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1218.387220] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1218.387402] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1218.387582] env[61986]: INFO nova.compute.manager [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1218.387851] env[61986]: DEBUG oslo.service.loopingcall [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1218.388054] env[61986]: DEBUG nova.compute.manager [-] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1218.388154] env[61986]: DEBUG nova.network.neutron [-] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1218.685999] env[61986]: DEBUG nova.compute.manager [req-e194138f-9aad-4d9e-9bbd-51bc77af9261 req-bc8c12bc-fffb-491f-b429-8fef425cc04f service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Received event network-vif-deleted-376c7f76-71f0-4f6f-8514-0769daaecb5b {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1218.686232] env[61986]: INFO nova.compute.manager [req-e194138f-9aad-4d9e-9bbd-51bc77af9261 req-bc8c12bc-fffb-491f-b429-8fef425cc04f service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Neutron deleted interface 376c7f76-71f0-4f6f-8514-0769daaecb5b; detaching it from the instance and deleting it from the info cache [ 1218.686412] env[61986]: DEBUG nova.network.neutron [req-e194138f-9aad-4d9e-9bbd-51bc77af9261 req-bc8c12bc-fffb-491f-b429-8fef425cc04f service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.794431] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.201s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.796666] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.927s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.842504] env[61986]: INFO nova.network.neutron [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating port 37976b99-2eee-45d3-93da-ce6f9f9813e2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1218.890778] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updating instance_info_cache with network_info: [{"id": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "address": "fa:16:3e:e3:03:02", "network": {"id": "64918e2a-576d-4390-8a9c-c4d5d69bd91a", "bridge": "br-int", "label": "tempest-ServersTestJSON-928860175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3d92ccda87d241068595992a1b8d3029", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a7188cb6-c9d7-449a-9362-ddb61a31feaf", "external-id": "nsx-vlan-transportzone-99", "segmentation_id": 99, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73c2047a-04", "ovs_interfaceid": "73c2047a-043c-4e3a-ae94-b9749f2229f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.156633] env[61986]: DEBUG nova.network.neutron [-] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.188981] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a37bc42f-db5c-4a89-8139-b6b712cac586 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.198258] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f109eed4-f00f-4dac-bb4a-4a5fba013b3c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.223666] env[61986]: DEBUG nova.compute.manager [req-e194138f-9aad-4d9e-9bbd-51bc77af9261 req-bc8c12bc-fffb-491f-b429-8fef425cc04f service nova] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Detach interface failed, port_id=376c7f76-71f0-4f6f-8514-0769daaecb5b, reason: Instance 3fb03cbe-3c13-4fce-8556-aaec6030b264 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1219.395041] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.395041] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1219.395041] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.395041] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.395041] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.395041] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.402553] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6600d044-0ee0-4b71-b9d7-6ca6d8514873 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.410291] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74296bb4-de75-4da0-ac2c-a45bb49c5a09 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.442236] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b789738-7364-44ac-ab26-02ae5e2ac1ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.449763] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3818f9-405d-44d7-a6ce-8477324d230c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.462987] env[61986]: DEBUG nova.compute.provider_tree [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.659754] env[61986]: INFO nova.compute.manager [-] [instance: 3fb03cbe-3c13-4fce-8556-aaec6030b264] Took 1.27 seconds to deallocate network for instance. [ 1219.965921] env[61986]: DEBUG nova.scheduler.client.report [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1220.166149] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.346617] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.346825] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.347064] env[61986]: DEBUG nova.network.neutron [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1220.711480] env[61986]: DEBUG nova.compute.manager [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-vif-plugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1220.712112] env[61986]: DEBUG oslo_concurrency.lockutils [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.712338] env[61986]: DEBUG oslo_concurrency.lockutils [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.712514] env[61986]: DEBUG oslo_concurrency.lockutils [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.712701] env[61986]: DEBUG nova.compute.manager [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] No waiting events found dispatching network-vif-plugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1220.712868] env[61986]: WARNING nova.compute.manager [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received unexpected event network-vif-plugged-37976b99-2eee-45d3-93da-ce6f9f9813e2 for instance with vm_state shelved_offloaded and task_state spawning. [ 1220.713041] env[61986]: DEBUG nova.compute.manager [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1220.713207] env[61986]: DEBUG nova.compute.manager [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing instance network info cache due to event network-changed-37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1220.713379] env[61986]: DEBUG oslo_concurrency.lockutils [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] Acquiring lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.977115] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.180s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.979933] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.814s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.980216] env[61986]: DEBUG nova.objects.instance [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid 3fb03cbe-3c13-4fce-8556-aaec6030b264 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.055260] env[61986]: DEBUG nova.network.neutron [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.540273] env[61986]: INFO nova.scheduler.client.report [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocation for migration e913fad5-a9ea-4133-8fb9-e06a6e08c666 [ 1221.557285] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.560326] env[61986]: DEBUG oslo_concurrency.lockutils [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] Acquired lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.560326] env[61986]: DEBUG nova.network.neutron [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Refreshing network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1221.582422] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3ed5caaf156db447f19639de1751786e',container_format='bare',created_at=2024-10-07T06:49:07Z,direct_url=,disk_format='vmdk',id=3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1757573630-shelved',owner='c88c612fe00a4ad589db67f3016dace6',properties=ImageMetaProps,protected=,size=31663616,status='active',tags=,updated_at=2024-10-07T06:49:20Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1221.582678] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1221.582874] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1221.583098] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1221.583254] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1221.583404] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1221.583613] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1221.583891] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1221.584115] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1221.584291] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1221.584470] env[61986]: DEBUG nova.virt.hardware [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1221.585662] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32690bdf-8481-4cd5-8e26-b7ff10e6d976 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.596355] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1fcf15-1f0b-41f7-a1ac-fbd266b76b41 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.601299] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d221e7c9-5156-4042-9cc5-643b34a8eff0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.612984] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:80:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37976b99-2eee-45d3-93da-ce6f9f9813e2', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1221.620192] env[61986]: DEBUG oslo.service.loopingcall [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1221.622163] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1221.622411] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5cdf9453-7535-4cfc-bc32-b6e40ab25feb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.636704] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f208251-81d9-4fe6-b12d-a6cd5108c3d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.670327] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1221.670327] env[61986]: value = "task-1160662" [ 1221.670327] env[61986]: _type = "Task" [ 1221.670327] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.671722] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e758635-ed5b-4de4-97af-01866c3656b6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.685570] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcf5c7c-00a0-45a2-b39c-27c09befe99a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.688289] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160662, 'name': CreateVM_Task} progress is 15%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.699581] env[61986]: DEBUG nova.compute.provider_tree [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.048663] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.488s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.181648] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160662, 'name': CreateVM_Task, 'duration_secs': 0.281038} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.181968] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1222.182747] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.182998] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.183391] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1222.183648] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5e69991-f64d-4b20-9fad-8079c3e2834d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.188722] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1222.188722] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336906-eae4-9df8-7bd5-17eda0d1d852" [ 1222.188722] env[61986]: _type = "Task" [ 1222.188722] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.196350] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52336906-eae4-9df8-7bd5-17eda0d1d852, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.202595] env[61986]: DEBUG nova.scheduler.client.report [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1222.269325] env[61986]: DEBUG nova.network.neutron [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updated VIF entry in instance network info cache for port 37976b99-2eee-45d3-93da-ce6f9f9813e2. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1222.269692] env[61986]: DEBUG nova.network.neutron [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [{"id": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "address": "fa:16:3e:3a:80:5d", "network": {"id": "8fbc0452-f3cf-4f45-a77c-8d72c46bad04", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-246799515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c88c612fe00a4ad589db67f3016dace6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37976b99-2e", "ovs_interfaceid": "37976b99-2eee-45d3-93da-ce6f9f9813e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.698870] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.699199] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Processing image 3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1222.699447] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.699599] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.699781] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1222.700041] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19aa2682-4b9b-45b5-afc6-d99ad51e656a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.707862] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1222.708060] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1222.709185] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.711075] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8e9f290-2b25-469e-b497-af282e38ed1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.716566] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1222.716566] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c819b3-d2e5-2191-b0d7-a138502b1c89" [ 1222.716566] env[61986]: _type = "Task" [ 1222.716566] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.723564] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c819b3-d2e5-2191-b0d7-a138502b1c89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.727626] env[61986]: INFO nova.scheduler.client.report [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance 3fb03cbe-3c13-4fce-8556-aaec6030b264 [ 1222.772387] env[61986]: DEBUG oslo_concurrency.lockutils [req-7a3f415c-0675-4564-8959-3e249accf90e req-5f7ba729-14e8-432f-b90b-674df87f1c48 service nova] Releasing lock "refresh_cache-c5a4fd56-25e2-4a95-8d2a-58361ef65e86" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.070957] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.071352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.071491] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.071735] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.071928] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.074186] env[61986]: INFO nova.compute.manager [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Terminating instance [ 1223.076008] env[61986]: DEBUG nova.compute.manager [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1223.076229] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1223.077098] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ada88b9-b97d-4ca6-b57c-94be69d6e941 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.084934] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1223.085176] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a5bd13d-759d-4655-bf63-d39d8f4b671b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.091064] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1223.091064] env[61986]: value = "task-1160663" [ 1223.091064] env[61986]: _type = "Task" [ 1223.091064] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.098439] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.226671] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Preparing fetch location {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1223.226927] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Fetch image to [datastore1] OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb/OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb.vmdk {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1223.227239] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Downloading stream optimized image 3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 to [datastore1] OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb/OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb.vmdk on the data store datastore1 as vApp {{(pid=61986) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1223.227407] env[61986]: DEBUG nova.virt.vmwareapi.images [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Downloading image file data 3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 to the ESX as VM named 'OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb' {{(pid=61986) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1223.235665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f5a4c3af-3440-4a29-a86b-898da441e266 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "3fb03cbe-3c13-4fce-8556-aaec6030b264" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.444s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.295321] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1223.295321] env[61986]: value = "resgroup-9" [ 1223.295321] env[61986]: _type = "ResourcePool" [ 1223.295321] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1223.295626] env[61986]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-84d26941-6ecf-42e2-af7a-6edd8cb701f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.317680] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease: (returnval){ [ 1223.317680] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234d8d1-6584-9f59-03e3-8bac2dc132b8" [ 1223.317680] env[61986]: _type = "HttpNfcLease" [ 1223.317680] env[61986]: } obtained for vApp import into resource pool (val){ [ 1223.317680] env[61986]: value = "resgroup-9" [ 1223.317680] env[61986]: _type = "ResourcePool" [ 1223.317680] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1223.317954] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the lease: (returnval){ [ 1223.317954] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234d8d1-6584-9f59-03e3-8bac2dc132b8" [ 1223.317954] env[61986]: _type = "HttpNfcLease" [ 1223.317954] env[61986]: } to be ready. {{(pid=61986) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1223.323669] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1223.323669] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234d8d1-6584-9f59-03e3-8bac2dc132b8" [ 1223.323669] env[61986]: _type = "HttpNfcLease" [ 1223.323669] env[61986]: } is initializing. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1223.601103] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160663, 'name': PowerOffVM_Task, 'duration_secs': 0.159679} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.601446] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.601663] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1223.601967] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8613201e-4082-4d18-a446-7074b37ffc36 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.661059] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1223.661337] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1223.661573] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleting the datastore file [datastore2] cff8354d-1eb9-446c-8c63-6697264c373f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1223.661909] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b679106-6716-431e-aa27-5267b81ec7d0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.668223] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for the task: (returnval){ [ 1223.668223] env[61986]: value = "task-1160666" [ 1223.668223] env[61986]: _type = "Task" [ 1223.668223] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.676926] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160666, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.825969] env[61986]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1223.825969] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234d8d1-6584-9f59-03e3-8bac2dc132b8" [ 1223.825969] env[61986]: _type = "HttpNfcLease" [ 1223.825969] env[61986]: } is ready. {{(pid=61986) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1223.826311] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1223.826311] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5234d8d1-6584-9f59-03e3-8bac2dc132b8" [ 1223.826311] env[61986]: _type = "HttpNfcLease" [ 1223.826311] env[61986]: }. {{(pid=61986) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1223.827077] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b91da1e-649a-40d2-8c3d-8ae9e6e13fc9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.833982] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6b6ee-e91f-b386-8e82-2dc2f6880b4c/disk-0.vmdk from lease info. {{(pid=61986) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1223.834177] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating HTTP connection to write to file with size = 31663616 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6b6ee-e91f-b386-8e82-2dc2f6880b4c/disk-0.vmdk. {{(pid=61986) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1223.898958] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c7dccd79-c083-498d-8466-8a3289e976c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.182882] env[61986]: DEBUG oslo_vmware.api [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Task: {'id': task-1160666, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130376} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.185390] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1224.185604] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1224.185792] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1224.186036] env[61986]: INFO nova.compute.manager [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1224.186333] env[61986]: DEBUG oslo.service.loopingcall [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1224.186908] env[61986]: DEBUG nova.compute.manager [-] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1224.187051] env[61986]: DEBUG nova.network.neutron [-] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1224.471653] env[61986]: DEBUG nova.compute.manager [req-f08e8b38-0a64-4a2c-ad89-fd268c0ff212 req-6d9268b4-a117-4a08-a862-762da6341f28 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Received event network-vif-deleted-9f357ac4-99e2-44ce-8f2c-3de7619c93fa {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1224.471944] env[61986]: INFO nova.compute.manager [req-f08e8b38-0a64-4a2c-ad89-fd268c0ff212 req-6d9268b4-a117-4a08-a862-762da6341f28 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Neutron deleted interface 9f357ac4-99e2-44ce-8f2c-3de7619c93fa; detaching it from the instance and deleting it from the info cache [ 1224.472165] env[61986]: DEBUG nova.network.neutron [req-f08e8b38-0a64-4a2c-ad89-fd268c0ff212 req-6d9268b4-a117-4a08-a862-762da6341f28 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.611905] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.612267] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.612515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.612710] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.612887] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.616067] env[61986]: INFO nova.compute.manager [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Terminating instance [ 1224.618236] env[61986]: DEBUG nova.compute.manager [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1224.618440] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1224.619380] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2c7666-7340-4304-a2cf-9f0a14b52c69 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.629168] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1224.630690] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e652411-0dd2-47c5-a510-00d3a742096b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.636881] env[61986]: DEBUG oslo_vmware.api [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1224.636881] env[61986]: value = "task-1160667" [ 1224.636881] env[61986]: _type = "Task" [ 1224.636881] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.645816] env[61986]: DEBUG oslo_vmware.api [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160667, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.949333] env[61986]: DEBUG nova.network.neutron [-] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.974970] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a85cb60b-3f8e-4acf-8988-493a130d2f7e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.985075] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b3cb39-a88b-47db-b721-eea7b7b49fcb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.003346] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Completed reading data from the image iterator. {{(pid=61986) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1225.003607] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6b6ee-e91f-b386-8e82-2dc2f6880b4c/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1225.004515] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadc6bb0-6f39-44ee-a700-d1aa00444149 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.010767] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6b6ee-e91f-b386-8e82-2dc2f6880b4c/disk-0.vmdk is in state: ready. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1225.011273] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6b6ee-e91f-b386-8e82-2dc2f6880b4c/disk-0.vmdk. {{(pid=61986) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1225.017751] env[61986]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-741df1d7-e097-4f11-99dd-1a5112c66dc5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.019365] env[61986]: DEBUG nova.compute.manager [req-f08e8b38-0a64-4a2c-ad89-fd268c0ff212 req-6d9268b4-a117-4a08-a862-762da6341f28 service nova] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Detach interface failed, port_id=9f357ac4-99e2-44ce-8f2c-3de7619c93fa, reason: Instance cff8354d-1eb9-446c-8c63-6697264c373f could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1225.147484] env[61986]: DEBUG oslo_vmware.api [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160667, 'name': PowerOffVM_Task, 'duration_secs': 0.185299} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.147755] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1225.147927] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1225.148196] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c402ab24-cfff-400e-8fc3-033d4ddd4a3e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.195684] env[61986]: DEBUG oslo_vmware.rw_handles [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c6b6ee-e91f-b386-8e82-2dc2f6880b4c/disk-0.vmdk. {{(pid=61986) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1225.196060] env[61986]: INFO nova.virt.vmwareapi.images [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Downloaded image file data 3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 [ 1225.196736] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a95f2f9-67d0-4cb3-8d9c-4d751ff52197 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.213355] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e77a42df-b004-4b8f-aa8b-b696949a6076 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.214991] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1225.215212] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1225.215395] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleting the datastore file [datastore1] 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1225.215633] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-963db3f4-7b12-41b1-ae85-0e2bf6c1ca85 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.221454] env[61986]: DEBUG oslo_vmware.api [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for the task: (returnval){ [ 1225.221454] env[61986]: value = "task-1160670" [ 1225.221454] env[61986]: _type = "Task" [ 1225.221454] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.228963] env[61986]: DEBUG oslo_vmware.api [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.240742] env[61986]: INFO nova.virt.vmwareapi.images [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] The imported VM was unregistered [ 1225.243284] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Caching image {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1225.243508] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Creating directory with path [datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1225.243762] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c83c6474-e65f-435d-bce5-d02c2dd63567 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.285663] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Created directory with path [datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268 {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1225.285972] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb/OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb.vmdk to [datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk. {{(pid=61986) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1225.286267] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-60726790-7160-469c-a42e-22d2932f99e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.293135] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1225.293135] env[61986]: value = "task-1160671" [ 1225.293135] env[61986]: _type = "Task" [ 1225.293135] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.301276] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.456344] env[61986]: INFO nova.compute.manager [-] [instance: cff8354d-1eb9-446c-8c63-6697264c373f] Took 1.27 seconds to deallocate network for instance. [ 1225.731240] env[61986]: DEBUG oslo_vmware.api [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Task: {'id': task-1160670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.475787} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.731587] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1225.731708] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1225.731922] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1225.732149] env[61986]: INFO nova.compute.manager [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1225.732414] env[61986]: DEBUG oslo.service.loopingcall [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1225.732622] env[61986]: DEBUG nova.compute.manager [-] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1225.732736] env[61986]: DEBUG nova.network.neutron [-] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1225.806498] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.963390] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.963665] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.963862] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.994723] env[61986]: INFO nova.scheduler.client.report [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Deleted allocations for instance cff8354d-1eb9-446c-8c63-6697264c373f [ 1226.306567] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.499763] env[61986]: DEBUG nova.compute.manager [req-5ff9bfb6-4602-4099-8208-d1fd732d6896 req-758fc12c-4bfa-4c61-b2cd-8f758df7f429 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Received event network-vif-deleted-73c2047a-043c-4e3a-ae94-b9749f2229f3 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1226.500026] env[61986]: INFO nova.compute.manager [req-5ff9bfb6-4602-4099-8208-d1fd732d6896 req-758fc12c-4bfa-4c61-b2cd-8f758df7f429 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Neutron deleted interface 73c2047a-043c-4e3a-ae94-b9749f2229f3; detaching it from the instance and deleting it from the info cache [ 1226.500173] env[61986]: DEBUG nova.network.neutron [req-5ff9bfb6-4602-4099-8208-d1fd732d6896 req-758fc12c-4bfa-4c61-b2cd-8f758df7f429 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.508164] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e85bf0fe-6520-4755-b966-65f77fdfce56 tempest-DeleteServersTestJSON-748295712 tempest-DeleteServersTestJSON-748295712-project-member] Lock "cff8354d-1eb9-446c-8c63-6697264c373f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.437s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.509744] env[61986]: DEBUG nova.network.neutron [-] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.806155] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.003748] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d81dfa9-d343-4930-8dbd-9f31e513f3c5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.012509] env[61986]: INFO nova.compute.manager [-] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Took 1.28 seconds to deallocate network for instance. [ 1227.018345] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7597a7-9ed6-4200-85d3-e864e0f9e99d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.049938] env[61986]: DEBUG nova.compute.manager [req-5ff9bfb6-4602-4099-8208-d1fd732d6896 req-758fc12c-4bfa-4c61-b2cd-8f758df7f429 service nova] [instance: 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350] Detach interface failed, port_id=73c2047a-043c-4e3a-ae94-b9749f2229f3, reason: Instance 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1227.307796] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.534420] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.534706] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.535048] env[61986]: DEBUG nova.objects.instance [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lazy-loading 'resources' on Instance uuid 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1227.806098] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.103860] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdfd236-6857-454a-b698-5057480c8f4f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.110982] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf0b204-f0e8-4c68-947a-866915c96814 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.139947] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedb66e3-7530-4f50-893b-87d2a081e112 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.146698] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c64c56-6780-4446-9318-e85e12195f4a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.160647] env[61986]: DEBUG nova.compute.provider_tree [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1228.306892] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160671, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.598657} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.307190] env[61986]: INFO nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb/OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb.vmdk to [datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk. [ 1228.307400] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Cleaning up location [datastore1] OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1228.307600] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ed6ea379-8a94-489b-b70d-735babefb9bb {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.307867] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7434360b-70ef-4ccc-9653-2a3b21edf00c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.315232] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1228.315232] env[61986]: value = "task-1160673" [ 1228.315232] env[61986]: _type = "Task" [ 1228.315232] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.323176] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.663468] env[61986]: DEBUG nova.scheduler.client.report [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1228.825670] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033458} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.825930] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1228.826120] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1228.826373] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk to [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1228.826624] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f46c148-419d-4b13-b063-82ca6dfd688a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.833359] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1228.833359] env[61986]: value = "task-1160674" [ 1228.833359] env[61986]: _type = "Task" [ 1228.833359] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.840728] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160674, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.168302] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.188686] env[61986]: INFO nova.scheduler.client.report [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Deleted allocations for instance 487d0b5e-27e6-4e9c-87b1-81e0aa9dd350 [ 1229.342952] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160674, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.698525] env[61986]: DEBUG oslo_concurrency.lockutils [None req-28942b1e-770c-40cc-9df4-269d3bf1ee07 tempest-ServersTestJSON-1386739338 tempest-ServersTestJSON-1386739338-project-member] Lock "487d0b5e-27e6-4e9c-87b1-81e0aa9dd350" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.086s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.846586] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160674, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.347622] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160674, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.848566] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160674, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.347136] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160674, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.166595} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.347433] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268/3d5b5fb5-69d0-4afd-a2cc-ae0dfceb8268.vmdk to [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1231.348258] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c890804-4627-40e8-89d3-976e7c746dbe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.374557] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1231.374557] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e57600c-173c-4860-9f76-fec046990a4b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.393694] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1231.393694] env[61986]: value = "task-1160675" [ 1231.393694] env[61986]: _type = "Task" [ 1231.393694] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.403649] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160675, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.903380] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160675, 'name': ReconfigVM_Task, 'duration_secs': 0.267862} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.903686] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Reconfigured VM instance instance-0000006d to attach disk [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86/c5a4fd56-25e2-4a95-8d2a-58361ef65e86.vmdk or device None with type streamOptimized {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1231.904416] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5b0938c-de80-40d0-8f64-772eb7339372 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.910444] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1231.910444] env[61986]: value = "task-1160676" [ 1231.910444] env[61986]: _type = "Task" [ 1231.910444] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.420683] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160676, 'name': Rename_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.922819] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160676, 'name': Rename_Task, 'duration_secs': 0.830346} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.923138] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1232.923452] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d9c9ba0-cdb4-41db-83ec-d43166a479b3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.930435] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1232.930435] env[61986]: value = "task-1160677" [ 1232.930435] env[61986]: _type = "Task" [ 1232.930435] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.939533] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.439457] env[61986]: DEBUG oslo_vmware.api [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160677, 'name': PowerOnVM_Task, 'duration_secs': 0.401199} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.439746] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1233.537414] env[61986]: DEBUG nova.compute.manager [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1233.538408] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57536fb-21e7-483d-a936-40640c6ab9f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.057031] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b2d839e5-045a-41ac-a651-070da0bd8167 tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.738s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.860224] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.860509] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.236563] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.236931] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.237081] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.237252] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.237428] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.239644] env[61986]: INFO nova.compute.manager [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Terminating instance [ 1235.241463] env[61986]: DEBUG nova.compute.manager [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1235.241660] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1235.242524] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde83b9d-d769-465a-aa69-4fd29b6e58c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.250163] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1235.250392] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fd11929-54ca-4633-ae3f-34e1b8d79924 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.256342] env[61986]: DEBUG oslo_vmware.api [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1235.256342] env[61986]: value = "task-1160678" [ 1235.256342] env[61986]: _type = "Task" [ 1235.256342] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.265233] env[61986]: DEBUG oslo_vmware.api [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.364662] env[61986]: DEBUG nova.compute.utils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1235.766982] env[61986]: DEBUG oslo_vmware.api [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160678, 'name': PowerOffVM_Task, 'duration_secs': 0.176618} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.767273] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1235.767446] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1235.767701] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d35efa75-cf24-4a06-8b0f-cab7e66c0b0d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.825508] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1235.825763] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1235.825910] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleting the datastore file [datastore1] c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.826217] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3627d7ba-babc-4153-bd52-325563699fba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.832346] env[61986]: DEBUG oslo_vmware.api [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for the task: (returnval){ [ 1235.832346] env[61986]: value = "task-1160680" [ 1235.832346] env[61986]: _type = "Task" [ 1235.832346] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.839558] env[61986]: DEBUG oslo_vmware.api [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.868355] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.342563] env[61986]: DEBUG oslo_vmware.api [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Task: {'id': task-1160680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129823} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.343031] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1236.343031] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1236.343263] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1236.343390] env[61986]: INFO nova.compute.manager [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1236.343640] env[61986]: DEBUG oslo.service.loopingcall [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1236.343832] env[61986]: DEBUG nova.compute.manager [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1236.343928] env[61986]: DEBUG nova.network.neutron [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1236.774271] env[61986]: DEBUG nova.compute.manager [req-ad54d431-9339-4078-ae0b-51a78beb83b2 req-d027228c-590e-43bb-b5c7-edb89ccc7ace service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Received event network-vif-deleted-37976b99-2eee-45d3-93da-ce6f9f9813e2 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1236.774480] env[61986]: INFO nova.compute.manager [req-ad54d431-9339-4078-ae0b-51a78beb83b2 req-d027228c-590e-43bb-b5c7-edb89ccc7ace service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Neutron deleted interface 37976b99-2eee-45d3-93da-ce6f9f9813e2; detaching it from the instance and deleting it from the info cache [ 1236.774668] env[61986]: DEBUG nova.network.neutron [req-ad54d431-9339-4078-ae0b-51a78beb83b2 req-d027228c-590e-43bb-b5c7-edb89ccc7ace service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.927755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.928118] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.928259] env[61986]: INFO nova.compute.manager [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Attaching volume 142a9778-51a4-4fbf-9913-e4ba0e7ff0dd to /dev/sdb [ 1236.969019] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eced8a0a-fa11-4da2-90af-6b897322d336 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.974736] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aaa42b0-e0af-474f-ae01-d3dcb0884792 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.989569] env[61986]: DEBUG nova.virt.block_device [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating existing volume attachment record: 7cdbd1ef-c6f0-4f65-8059-95930c7e362e {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1237.254892] env[61986]: DEBUG nova.network.neutron [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.277664] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d135437-0b17-4250-ac30-e170fabb3646 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.286389] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa7376c-19e6-43f1-9b2d-b50a60e16310 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.311416] env[61986]: DEBUG nova.compute.manager [req-ad54d431-9339-4078-ae0b-51a78beb83b2 req-d027228c-590e-43bb-b5c7-edb89ccc7ace service nova] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Detach interface failed, port_id=37976b99-2eee-45d3-93da-ce6f9f9813e2, reason: Instance c5a4fd56-25e2-4a95-8d2a-58361ef65e86 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1237.757444] env[61986]: INFO nova.compute.manager [-] [instance: c5a4fd56-25e2-4a95-8d2a-58361ef65e86] Took 1.41 seconds to deallocate network for instance. [ 1238.265688] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.266060] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.266060] env[61986]: DEBUG nova.objects.instance [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lazy-loading 'resources' on Instance uuid c5a4fd56-25e2-4a95-8d2a-58361ef65e86 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.826187] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f184b1-258f-4c5e-8929-2a3a11016d5e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.833432] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201d4bc9-bc98-434f-a209-e92bcee6c101 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.863187] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e3d53f-c581-499c-80b2-9c157aa6b4fa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.869693] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892cdb18-a94e-4e8f-b337-22345cd83370 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.882247] env[61986]: DEBUG nova.compute.provider_tree [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.385202] env[61986]: DEBUG nova.scheduler.client.report [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1239.890466] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.911518] env[61986]: INFO nova.scheduler.client.report [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Deleted allocations for instance c5a4fd56-25e2-4a95-8d2a-58361ef65e86 [ 1240.419968] env[61986]: DEBUG oslo_concurrency.lockutils [None req-9f631a0b-8cb7-44f1-9a4e-7feb9d3d271e tempest-ServerActionsTestOtherB-49737381 tempest-ServerActionsTestOtherB-49737381-project-member] Lock "c5a4fd56-25e2-4a95-8d2a-58361ef65e86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.183s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.538850] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1241.539123] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1241.540071] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba39d9b-d22d-4f56-b6d0-d3270b86d115 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.556430] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55dcee9c-2104-4280-9082-b4a4d8aec067 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.579778] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd/volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1241.579975] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5930d694-7999-461f-a878-7164c3613e70 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.596721] env[61986]: DEBUG oslo_vmware.api [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1241.596721] env[61986]: value = "task-1160686" [ 1241.596721] env[61986]: _type = "Task" [ 1241.596721] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.603968] env[61986]: DEBUG oslo_vmware.api [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.106651] env[61986]: DEBUG oslo_vmware.api [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160686, 'name': ReconfigVM_Task, 'duration_secs': 0.450645} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.106917] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfigured VM instance instance-00000073 to attach disk [datastore2] volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd/volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1242.111544] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a118c2ad-e128-42f6-804d-43cea34fb751 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.127162] env[61986]: DEBUG oslo_vmware.api [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1242.127162] env[61986]: value = "task-1160687" [ 1242.127162] env[61986]: _type = "Task" [ 1242.127162] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.134490] env[61986]: DEBUG oslo_vmware.api [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.637569] env[61986]: DEBUG oslo_vmware.api [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160687, 'name': ReconfigVM_Task, 'duration_secs': 0.151914} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.637915] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1243.672754] env[61986]: DEBUG nova.objects.instance [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'flavor' on Instance uuid 66267549-1945-4c63-8b3b-c090a4ea1a28 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.180640] env[61986]: DEBUG oslo_concurrency.lockutils [None req-3063f989-7b88-4759-9333-3035ca06930d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.251s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.630429] env[61986]: INFO nova.compute.manager [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Rebuilding instance [ 1244.697758] env[61986]: DEBUG nova.compute.manager [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1244.698709] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a21a6a-6743-4c0f-8d46-a51289f43287 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.210922] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1245.211272] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2dd8a4c1-268f-43ad-851a-0f58afb09f64 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.217916] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1245.217916] env[61986]: value = "task-1160688" [ 1245.217916] env[61986]: _type = "Task" [ 1245.217916] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.225897] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.727702] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160688, 'name': PowerOffVM_Task, 'duration_secs': 0.208132} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.728066] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1245.784518] env[61986]: INFO nova.compute.manager [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Detaching volume 142a9778-51a4-4fbf-9913-e4ba0e7ff0dd [ 1245.813674] env[61986]: INFO nova.virt.block_device [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Attempting to driver detach volume 142a9778-51a4-4fbf-9913-e4ba0e7ff0dd from mountpoint /dev/sdb [ 1245.813906] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1245.814155] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1245.814997] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7921f90f-d676-486e-a291-9bc586eae01b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.836413] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fda748-1159-40ec-bd17-7e5d6b738010 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.842796] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f84927-0e6e-42d3-a90e-b4a2d7dbdace {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.862084] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21e98b3-bda9-4a0e-b7ca-02962c8dbe00 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.875560] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] The volume has not been displaced from its original location: [datastore2] volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd/volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1245.880635] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1245.880878] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d8c9920-26c3-4e9e-9f68-e5b4dbb2dbbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.898885] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1245.898885] env[61986]: value = "task-1160689" [ 1245.898885] env[61986]: _type = "Task" [ 1245.898885] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.907289] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160689, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.410938] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160689, 'name': ReconfigVM_Task, 'duration_secs': 0.279157} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.411364] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1246.416043] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b03bf238-580e-4e12-b8e5-19ba5cdd9a90 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.432044] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1246.432044] env[61986]: value = "task-1160690" [ 1246.432044] env[61986]: _type = "Task" [ 1246.432044] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.439648] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160690, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.941064] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160690, 'name': ReconfigVM_Task, 'duration_secs': 0.249418} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.941421] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1247.992691] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1247.993091] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e01fcec4-9c5c-4ddb-8f40-49ac8634b836 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.000859] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1248.000859] env[61986]: value = "task-1160691" [ 1248.000859] env[61986]: _type = "Task" [ 1248.000859] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.009437] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1248.009647] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1248.009855] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1248.010581] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b830152-f7cf-4209-a5f3-2a9ff247a099 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.027974] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2327e22d-1c0f-492f-ab43-e7d7ee49edda {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.033233] env[61986]: WARNING nova.virt.vmwareapi.driver [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1248.033474] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1248.034158] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd3e87b-8164-49b8-8315-354acb0e648b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.039898] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1248.040122] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-319e62d5-ead1-425d-82a4-32af80a0da73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.041484] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1251.041848] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1251.041886] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore2] 66267549-1945-4c63-8b3b-c090a4ea1a28 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1251.042248] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bee5d9e6-932d-43af-9fee-c177cd43d9c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.048330] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1251.048330] env[61986]: value = "task-1160693" [ 1251.048330] env[61986]: _type = "Task" [ 1251.048330] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.055945] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160693, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.557595] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160693, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244886} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.557821] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1251.558040] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1251.558246] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1251.593085] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.593358] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.593570] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.593755] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.593922] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.595806] env[61986]: INFO nova.compute.manager [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Terminating instance [ 1251.597482] env[61986]: DEBUG nova.compute.manager [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1251.597674] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1251.598635] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abc4378-f053-413f-a546-d82f619557d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.605556] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1251.605779] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7169b6f-8fd8-4fbe-8d59-f75d24c066c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.612801] env[61986]: DEBUG oslo_vmware.api [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1251.612801] env[61986]: value = "task-1160694" [ 1251.612801] env[61986]: _type = "Task" [ 1251.612801] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.621422] env[61986]: DEBUG oslo_vmware.api [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.063923] env[61986]: INFO nova.virt.block_device [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Booting with volume 142a9778-51a4-4fbf-9913-e4ba0e7ff0dd at /dev/sdb [ 1252.098497] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08b692d3-d372-4bb7-a068-b5d151f43169 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.107477] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a92901-19bc-49fc-b96e-1ebb6254fe49 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.125372] env[61986]: DEBUG oslo_vmware.api [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160694, 'name': PowerOffVM_Task, 'duration_secs': 0.16847} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.125623] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1252.125798] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1252.126036] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee9482e9-427a-4a76-9a1d-5d9771ad1401 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.132589] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1ebf1f6-a3a0-4e2a-8d45-d19e6e6fbf6e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.139926] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19dc6dd0-5bc8-4ca6-9eff-6281f72e2af4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.161711] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae38bd4b-4c18-45f5-8814-b1644b4ad735 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.167307] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a50bcf0-8f8b-4a3f-9f4c-7ff5c8de7ba5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.180950] env[61986]: DEBUG nova.virt.block_device [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating existing volume attachment record: a7e946b8-4903-4b2a-a775-23d85911a955 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1252.184078] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1252.184285] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1252.184471] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleting the datastore file [datastore1] 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1252.184708] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7effb015-c703-4ff0-beeb-0e9c83344e5e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.189828] env[61986]: DEBUG oslo_vmware.api [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for the task: (returnval){ [ 1252.189828] env[61986]: value = "task-1160696" [ 1252.189828] env[61986]: _type = "Task" [ 1252.189828] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.197060] env[61986]: DEBUG oslo_vmware.api [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160696, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.699652] env[61986]: DEBUG oslo_vmware.api [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Task: {'id': task-1160696, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124063} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.699986] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1252.700144] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1252.700330] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1252.700508] env[61986]: INFO nova.compute.manager [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1252.700752] env[61986]: DEBUG oslo.service.loopingcall [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1252.700949] env[61986]: DEBUG nova.compute.manager [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1252.701059] env[61986]: DEBUG nova.network.neutron [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1253.416910] env[61986]: DEBUG nova.compute.manager [req-fadb58db-3d97-4acc-b3f9-4730c37540df req-aa7d6bb3-e2a4-45f7-922e-fe78dd0b5a33 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Received event network-vif-deleted-0b5285ba-76dd-4e56-aa09-5ac60d84235f {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1253.417172] env[61986]: INFO nova.compute.manager [req-fadb58db-3d97-4acc-b3f9-4730c37540df req-aa7d6bb3-e2a4-45f7-922e-fe78dd0b5a33 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Neutron deleted interface 0b5285ba-76dd-4e56-aa09-5ac60d84235f; detaching it from the instance and deleting it from the info cache [ 1253.417440] env[61986]: DEBUG nova.network.neutron [req-fadb58db-3d97-4acc-b3f9-4730c37540df req-aa7d6bb3-e2a4-45f7-922e-fe78dd0b5a33 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.894368] env[61986]: DEBUG nova.network.neutron [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.921055] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acc7e05c-ae90-4887-89b3-c188dfc289c2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.929800] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f31da0-00e2-498a-839f-ea362cf172cc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.954838] env[61986]: DEBUG nova.compute.manager [req-fadb58db-3d97-4acc-b3f9-4730c37540df req-aa7d6bb3-e2a4-45f7-922e-fe78dd0b5a33 service nova] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Detach interface failed, port_id=0b5285ba-76dd-4e56-aa09-5ac60d84235f, reason: Instance 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1254.301572] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1254.301848] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1254.302059] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1254.302275] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1254.302427] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1254.302577] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1254.302785] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1254.302956] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1254.303190] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1254.303368] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1254.303546] env[61986]: DEBUG nova.virt.hardware [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1254.304507] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df86c76-007c-45fe-9688-b5296922bfdb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.312336] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eba414-86cc-46ce-b7ac-dcbd0b1c2c08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.325667] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:92:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c570cf8-315d-46b2-a056-e00ad031a6ab', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1254.333091] env[61986]: DEBUG oslo.service.loopingcall [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1254.333332] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1254.333537] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4740e105-db25-44d2-a11f-7d0ae67b8490 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.352282] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1254.352282] env[61986]: value = "task-1160697" [ 1254.352282] env[61986]: _type = "Task" [ 1254.352282] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.360257] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160697, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.396920] env[61986]: INFO nova.compute.manager [-] [instance: 4ba294fa-5a23-45eb-95ed-ac41d541ebf7] Took 1.70 seconds to deallocate network for instance. [ 1254.862437] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160697, 'name': CreateVM_Task, 'duration_secs': 0.267921} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.862800] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1254.863278] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.863446] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.863770] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1254.864028] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1011e152-7c0b-45e5-84fe-c1c7f2b3118c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.868298] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1254.868298] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5272a4bd-3ab9-706e-99bc-c8aa78de70e3" [ 1254.868298] env[61986]: _type = "Task" [ 1254.868298] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.876835] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5272a4bd-3ab9-706e-99bc-c8aa78de70e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.902837] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.903073] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.903292] env[61986]: DEBUG nova.objects.instance [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lazy-loading 'resources' on Instance uuid 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1255.378924] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5272a4bd-3ab9-706e-99bc-c8aa78de70e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010548} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.379254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1255.379489] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1255.379720] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.379891] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.380064] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1255.380323] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-865ea4b7-72eb-4054-aaaa-67d3cb8b5134 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.387596] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1255.387778] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1255.388440] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46a90a23-8386-4ac8-a942-b6c4f97e3681 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.393201] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1255.393201] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526a0107-faed-0b4c-2638-7e771a9bdb52" [ 1255.393201] env[61986]: _type = "Task" [ 1255.393201] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.400271] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526a0107-faed-0b4c-2638-7e771a9bdb52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.457670] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fea3db-cc8c-4aa5-b106-6ad8eceaf1f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.464565] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0691a8c7-1a5b-438d-a517-73cc15037439 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.493547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888ced49-eab1-4a9e-b22e-e182d7d1713c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.499964] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e02f74a-35c2-4fa5-a78c-28b9a1287326 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.512338] env[61986]: DEBUG nova.compute.provider_tree [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1255.903690] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]526a0107-faed-0b4c-2638-7e771a9bdb52, 'name': SearchDatastore_Task, 'duration_secs': 0.007204} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.904427] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1712b75-55ee-4933-a32f-a460a2765fea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.909100] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1255.909100] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5205083c-0461-fe68-fb1a-cb2a9affb4e4" [ 1255.909100] env[61986]: _type = "Task" [ 1255.909100] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.917542] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5205083c-0461-fe68-fb1a-cb2a9affb4e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.015941] env[61986]: DEBUG nova.scheduler.client.report [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1256.419805] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5205083c-0461-fe68-fb1a-cb2a9affb4e4, 'name': SearchDatastore_Task, 'duration_secs': 0.008716} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.420100] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.420362] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1256.420626] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba572d1d-7e82-4211-aa5b-14137245f496 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.427249] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1256.427249] env[61986]: value = "task-1160698" [ 1256.427249] env[61986]: _type = "Task" [ 1256.427249] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.434154] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.521257] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.543752] env[61986]: INFO nova.scheduler.client.report [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Deleted allocations for instance 4ba294fa-5a23-45eb-95ed-ac41d541ebf7 [ 1256.937084] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.422637} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.937366] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore1] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1256.937585] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1256.937845] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d70c347-6dd4-46fc-a4d9-11f2de6127ae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.944378] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1256.944378] env[61986]: value = "task-1160699" [ 1256.944378] env[61986]: _type = "Task" [ 1256.944378] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.952519] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.051504] env[61986]: DEBUG oslo_concurrency.lockutils [None req-fbea51b9-e191-4dc9-8e29-760db65909b6 tempest-AttachVolumeShelveTestJSON-494191263 tempest-AttachVolumeShelveTestJSON-494191263-project-member] Lock "4ba294fa-5a23-45eb-95ed-ac41d541ebf7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.457s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.454407] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104586} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.454709] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1257.455422] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586ff8d4-534f-4538-9de8-21e7184108b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.477060] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1257.477326] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f9199d0-1599-4833-8fa3-129c56b0375e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.496495] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1257.496495] env[61986]: value = "task-1160700" [ 1257.496495] env[61986]: _type = "Task" [ 1257.496495] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.504109] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160700, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.006799] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160700, 'name': ReconfigVM_Task, 'duration_secs': 0.262743} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.007168] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfigured VM instance instance-00000073 to attach disk [datastore1] 66267549-1945-4c63-8b3b-c090a4ea1a28/66267549-1945-4c63-8b3b-c090a4ea1a28.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1258.008345] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'encrypted': False, 'size': 0, 'encryption_secret_uuid': None, 'encryption_format': None, 'boot_index': 0, 'disk_bus': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'encryption_options': None, 'image_id': '7b73cf0e-555e-4258-a7ba-6cecbc90b6c5'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'}, 'guest_format': None, 'attachment_id': 'a7e946b8-4903-4b2a-a775-23d85911a955', 'boot_index': None, 'disk_bus': None, 'mount_device': '/dev/sdb', 'device_type': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61986) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1258.008619] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1258.008871] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1258.009739] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293a6480-2c8e-4e88-a650-b53ce0f60cff {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.025498] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c48dd8b-6b81-4ae4-8439-4148b867ed5b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.051388] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd/volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1258.051686] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b07cc19f-2d53-4406-a34f-8a00f3861475 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.069221] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1258.069221] env[61986]: value = "task-1160701" [ 1258.069221] env[61986]: _type = "Task" [ 1258.069221] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.077091] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160701, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.578965] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160701, 'name': ReconfigVM_Task, 'duration_secs': 0.303878} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.579279] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfigured VM instance instance-00000073 to attach disk [datastore2] volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd/volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1258.583906] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af4f3c05-d041-44b2-ab68-d3292327d0ea {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.599012] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1258.599012] env[61986]: value = "task-1160703" [ 1258.599012] env[61986]: _type = "Task" [ 1258.599012] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.606256] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160703, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.109448] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160703, 'name': ReconfigVM_Task, 'duration_secs': 0.141161} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.109823] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1259.110334] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7ed5a9f-a0f0-4df1-8ec3-7b0528673175 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.116964] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1259.116964] env[61986]: value = "task-1160704" [ 1259.116964] env[61986]: _type = "Task" [ 1259.116964] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.124164] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160704, 'name': Rename_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.627302] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160704, 'name': Rename_Task, 'duration_secs': 0.134951} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.627590] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1259.627844] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c78a536-724c-408c-81b3-019e251d3961 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.633333] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1259.633333] env[61986]: value = "task-1160705" [ 1259.633333] env[61986]: _type = "Task" [ 1259.633333] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.640314] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160705, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.142644] env[61986]: DEBUG oslo_vmware.api [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160705, 'name': PowerOnVM_Task, 'duration_secs': 0.428459} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.142981] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1260.143246] env[61986]: DEBUG nova.compute.manager [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1260.144007] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cd5b53-1f66-48d3-8460-eb1199eadbb0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.659908] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1260.660244] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1260.660575] env[61986]: DEBUG nova.objects.instance [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61986) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1261.669048] env[61986]: DEBUG oslo_concurrency.lockutils [None req-48b67894-7504-4b80-b849-ada51b01e484 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.056796] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.057311] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.560395] env[61986]: INFO nova.compute.manager [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Detaching volume 142a9778-51a4-4fbf-9913-e4ba0e7ff0dd [ 1262.597150] env[61986]: INFO nova.virt.block_device [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Attempting to driver detach volume 142a9778-51a4-4fbf-9913-e4ba0e7ff0dd from mountpoint /dev/sdb [ 1262.597228] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1262.597456] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1262.598447] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbde09d-d512-4d25-b8e2-8ee1f17274e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.622722] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc9cd39-c6c1-4d81-85d7-b6d2a1e6f326 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.629920] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ef9289-e4d3-41e5-a299-9e7497b78d0e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.650195] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21fbcda-f61b-4c40-9228-34d4a88380db {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.664934] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] The volume has not been displaced from its original location: [datastore2] volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd/volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1262.670180] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1262.670574] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5e2a1c0-a5d8-47d0-81e9-6a7ca555b09d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.692713] env[61986]: DEBUG oslo_vmware.api [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1262.692713] env[61986]: value = "task-1160706" [ 1262.692713] env[61986]: _type = "Task" [ 1262.692713] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.700786] env[61986]: DEBUG oslo_vmware.api [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160706, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.202800] env[61986]: DEBUG oslo_vmware.api [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160706, 'name': ReconfigVM_Task, 'duration_secs': 0.235463} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.203190] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1263.207712] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-394764e8-0754-48b2-a323-a28b475851c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.222144] env[61986]: DEBUG oslo_vmware.api [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1263.222144] env[61986]: value = "task-1160707" [ 1263.222144] env[61986]: _type = "Task" [ 1263.222144] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.229351] env[61986]: DEBUG oslo_vmware.api [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160707, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.731919] env[61986]: DEBUG oslo_vmware.api [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160707, 'name': ReconfigVM_Task, 'duration_secs': 0.210338} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.732948] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252502', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'name': 'volume-142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66267549-1945-4c63-8b3b-c090a4ea1a28', 'attached_at': '', 'detached_at': '', 'volume_id': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd', 'serial': '142a9778-51a4-4fbf-9913-e4ba0e7ff0dd'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1264.277357] env[61986]: DEBUG nova.objects.instance [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'flavor' on Instance uuid 66267549-1945-4c63-8b3b-c090a4ea1a28 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.284726] env[61986]: DEBUG oslo_concurrency.lockutils [None req-0463a5a8-bf15-4859-8f35-4cde0d65be88 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.227s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.315072] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.315473] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.315553] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "66267549-1945-4c63-8b3b-c090a4ea1a28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.315786] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.315917] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.318900] env[61986]: INFO nova.compute.manager [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Terminating instance [ 1266.320868] env[61986]: DEBUG nova.compute.manager [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1266.321090] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1266.321971] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c13ee068-2468-45db-a1dc-68817704db31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.330181] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1266.330434] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6f78328-d63d-4194-9fbe-8e4892104e3b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.337667] env[61986]: DEBUG oslo_vmware.api [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1266.337667] env[61986]: value = "task-1160708" [ 1266.337667] env[61986]: _type = "Task" [ 1266.337667] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.345201] env[61986]: DEBUG oslo_vmware.api [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.846875] env[61986]: DEBUG oslo_vmware.api [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160708, 'name': PowerOffVM_Task, 'duration_secs': 0.185895} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.847223] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1266.847457] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1266.847745] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd5e9b16-2600-4d4a-8f7b-0f9b50c5caba {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.906178] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1266.906396] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1266.906583] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore1] 66267549-1945-4c63-8b3b-c090a4ea1a28 {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1266.906849] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2465572-5e1d-4d89-af6d-9c82c41d4f27 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.912961] env[61986]: DEBUG oslo_vmware.api [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1266.912961] env[61986]: value = "task-1160710" [ 1266.912961] env[61986]: _type = "Task" [ 1266.912961] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.920216] env[61986]: DEBUG oslo_vmware.api [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160710, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.421918] env[61986]: DEBUG oslo_vmware.api [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160710, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121038} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.422321] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1267.422435] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1267.422617] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1267.422794] env[61986]: INFO nova.compute.manager [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1267.423049] env[61986]: DEBUG oslo.service.loopingcall [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1267.423292] env[61986]: DEBUG nova.compute.manager [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1267.423395] env[61986]: DEBUG nova.network.neutron [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1267.842598] env[61986]: DEBUG nova.compute.manager [req-3dc15e11-bfe8-420b-bd67-074dd9d2aca0 req-6768b7b6-3d18-4515-b65f-70b8373bf69c service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Received event network-vif-deleted-0c570cf8-315d-46b2-a056-e00ad031a6ab {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1267.843132] env[61986]: INFO nova.compute.manager [req-3dc15e11-bfe8-420b-bd67-074dd9d2aca0 req-6768b7b6-3d18-4515-b65f-70b8373bf69c service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Neutron deleted interface 0c570cf8-315d-46b2-a056-e00ad031a6ab; detaching it from the instance and deleting it from the info cache [ 1267.843132] env[61986]: DEBUG nova.network.neutron [req-3dc15e11-bfe8-420b-bd67-074dd9d2aca0 req-6768b7b6-3d18-4515-b65f-70b8373bf69c service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.322438] env[61986]: DEBUG nova.network.neutron [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.345312] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be77f78e-f16a-483f-9870-9c81ab3795c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.355726] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0ef701-a9b6-46c3-8086-22ea83429a4a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.379451] env[61986]: DEBUG nova.compute.manager [req-3dc15e11-bfe8-420b-bd67-074dd9d2aca0 req-6768b7b6-3d18-4515-b65f-70b8373bf69c service nova] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Detach interface failed, port_id=0c570cf8-315d-46b2-a056-e00ad031a6ab, reason: Instance 66267549-1945-4c63-8b3b-c090a4ea1a28 could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1268.825524] env[61986]: INFO nova.compute.manager [-] [instance: 66267549-1945-4c63-8b3b-c090a4ea1a28] Took 1.40 seconds to deallocate network for instance. [ 1269.214034] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.214233] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1269.332792] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.333271] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.333634] env[61986]: DEBUG nova.objects.instance [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'resources' on Instance uuid 66267549-1945-4c63-8b3b-c090a4ea1a28 {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1269.877439] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93c0bfd-b6ed-4d5a-945a-9f4ee98b40c6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.884761] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b974d24-a59b-4ab0-b13b-0d5ce7943555 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.913942] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ec25a4-41b3-4458-93cb-45d549725962 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.920595] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752cfb77-9c0d-49c3-ad0a-b76a456ab012 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.933432] env[61986]: DEBUG nova.compute.provider_tree [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.214213] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1270.437036] env[61986]: DEBUG nova.scheduler.client.report [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1270.942727] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.961423] env[61986]: INFO nova.scheduler.client.report [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted allocations for instance 66267549-1945-4c63-8b3b-c090a4ea1a28 [ 1271.213850] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.213850] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1271.472139] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e89b194e-781f-4b84-9665-10278ac7abc8 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "66267549-1945-4c63-8b3b-c090a4ea1a28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.156s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.717094] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.717225] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.717352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.717508] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1271.718468] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1ba488-9522-49cc-aa4b-04546d1483ca {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.727034] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e272e9-7164-406a-a37a-442ac527d671 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.740356] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecfa9e0-acba-4ed2-aa35-4d5926405067 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.746547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91655fcc-1ad3-4818-9875-2bbf7f654ccc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.775242] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180751MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1271.775404] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.775600] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.798525] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a6094cc4-7d17-4858-abae-972425241c8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1272.887726] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.887967] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.301358] env[61986]: INFO nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec has allocations against this compute host but is not found in the database. [ 1273.301630] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1273.301726] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1273.340251] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c344c01-14ca-4ff7-a989-a7935ceeb244 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.347854] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac17e55-e5c0-4c05-845c-318abc9c1ec4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.377284] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973f57a7-84ec-448f-bcc0-0048c70a33bf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.384396] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97eb5308-3a41-4ad8-997a-fdfd4d6dbdae {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.396855] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1273.399420] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.904546] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1273.919015] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1274.411297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1274.411509] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.636s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.411796] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.493s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1274.413361] env[61986]: INFO nova.compute.claims [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1275.457235] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db9c1ae-62fe-47ec-a210-c9c6cbd4e18d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.465526] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720d0011-0279-4dfa-9385-03a184f0d019 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.495061] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4dd66b5-ab27-4a4d-a0d6-0e22300f32e1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.502015] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25260120-dc41-44d6-b307-08499d16d291 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.515682] env[61986]: DEBUG nova.compute.provider_tree [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1276.018746] env[61986]: DEBUG nova.scheduler.client.report [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1276.416200] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.416456] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.523373] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.523909] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1276.920971] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.921174] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1277.028556] env[61986]: DEBUG nova.compute.utils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1277.029986] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1277.030181] env[61986]: DEBUG nova.network.neutron [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1277.078441] env[61986]: DEBUG nova.policy [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c804063142764cac8244fd3d6bd71e16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838620f936dc4489be8b99ef87bf37ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1277.317730] env[61986]: DEBUG nova.network.neutron [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Successfully created port: 3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1277.463767] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1277.463914] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquired lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1277.464079] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Forcefully refreshing network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1277.534385] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1278.543856] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1278.574619] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1278.575038] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1278.575314] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1278.575625] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1278.575879] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1278.576158] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1278.576565] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1278.576778] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1278.577090] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1278.577379] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1278.577680] env[61986]: DEBUG nova.virt.hardware [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1278.579020] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70539091-044f-41a3-ad29-dcdf4eccf21c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.588587] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d61c53-d86f-429b-b956-d9158335e982 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.744309] env[61986]: DEBUG nova.compute.manager [req-f27f36d8-6422-4f94-99d5-b191fb62a9a5 req-a8b06fd5-ebe2-48e1-a108-3d32ef3776d7 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Received event network-vif-plugged-3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1278.744537] env[61986]: DEBUG oslo_concurrency.lockutils [req-f27f36d8-6422-4f94-99d5-b191fb62a9a5 req-a8b06fd5-ebe2-48e1-a108-3d32ef3776d7 service nova] Acquiring lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.744741] env[61986]: DEBUG oslo_concurrency.lockutils [req-f27f36d8-6422-4f94-99d5-b191fb62a9a5 req-a8b06fd5-ebe2-48e1-a108-3d32ef3776d7 service nova] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.744911] env[61986]: DEBUG oslo_concurrency.lockutils [req-f27f36d8-6422-4f94-99d5-b191fb62a9a5 req-a8b06fd5-ebe2-48e1-a108-3d32ef3776d7 service nova] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.745123] env[61986]: DEBUG nova.compute.manager [req-f27f36d8-6422-4f94-99d5-b191fb62a9a5 req-a8b06fd5-ebe2-48e1-a108-3d32ef3776d7 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] No waiting events found dispatching network-vif-plugged-3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1278.745291] env[61986]: WARNING nova.compute.manager [req-f27f36d8-6422-4f94-99d5-b191fb62a9a5 req-a8b06fd5-ebe2-48e1-a108-3d32ef3776d7 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Received unexpected event network-vif-plugged-3371718c-6628-4377-8a53-64b3ac1baf38 for instance with vm_state building and task_state spawning. [ 1278.798966] env[61986]: DEBUG nova.network.neutron [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [{"id": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "address": "fa:16:3e:22:d2:28", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5de45ae0-ff", "ovs_interfaceid": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1279.302032] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Releasing lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1279.302380] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updated the network info_cache for instance {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1279.302601] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.302771] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.302916] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.341863] env[61986]: DEBUG nova.network.neutron [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Successfully updated port: 3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1279.365399] env[61986]: DEBUG nova.compute.manager [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Received event network-changed-3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1279.365720] env[61986]: DEBUG nova.compute.manager [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Refreshing instance network info cache due to event network-changed-3371718c-6628-4377-8a53-64b3ac1baf38. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1279.365898] env[61986]: DEBUG oslo_concurrency.lockutils [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] Acquiring lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1279.366072] env[61986]: DEBUG oslo_concurrency.lockutils [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] Acquired lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.366278] env[61986]: DEBUG nova.network.neutron [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Refreshing network info cache for port 3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1279.844391] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1279.897491] env[61986]: DEBUG nova.network.neutron [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1279.966757] env[61986]: DEBUG nova.network.neutron [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1280.470989] env[61986]: DEBUG oslo_concurrency.lockutils [req-344b0875-a6aa-4967-8d64-6856ed46dd1a req-e58f960e-d34c-4f7f-9c65-0e6a693cc569 service nova] Releasing lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1280.471384] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.471550] env[61986]: DEBUG nova.network.neutron [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1281.001837] env[61986]: DEBUG nova.network.neutron [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1281.119711] env[61986]: DEBUG nova.network.neutron [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Updating instance_info_cache with network_info: [{"id": "3371718c-6628-4377-8a53-64b3ac1baf38", "address": "fa:16:3e:d7:4a:dd", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3371718c-66", "ovs_interfaceid": "3371718c-6628-4377-8a53-64b3ac1baf38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1281.621966] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1281.622361] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Instance network_info: |[{"id": "3371718c-6628-4377-8a53-64b3ac1baf38", "address": "fa:16:3e:d7:4a:dd", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3371718c-66", "ovs_interfaceid": "3371718c-6628-4377-8a53-64b3ac1baf38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1281.622813] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:4a:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3371718c-6628-4377-8a53-64b3ac1baf38', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1281.630891] env[61986]: DEBUG oslo.service.loopingcall [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1281.631114] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1281.631337] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be729c53-76cc-4dc2-bd46-973a5f329fbc {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.651940] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1281.651940] env[61986]: value = "task-1160711" [ 1281.651940] env[61986]: _type = "Task" [ 1281.651940] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.659514] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160711, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.161871] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160711, 'name': CreateVM_Task, 'duration_secs': 0.318505} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.162174] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1282.162864] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1282.162961] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.163318] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1282.163570] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1a9acd1-d569-4633-b5f8-2b9ba5019536 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.167934] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1282.167934] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5299337d-6f9f-c69c-a649-6986cc5668de" [ 1282.167934] env[61986]: _type = "Task" [ 1282.167934] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.176682] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5299337d-6f9f-c69c-a649-6986cc5668de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.677976] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5299337d-6f9f-c69c-a649-6986cc5668de, 'name': SearchDatastore_Task, 'duration_secs': 0.011341} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.678254] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.678489] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1282.678722] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1282.678872] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.679073] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1282.679338] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8959730b-724c-4e0f-98db-8fd46a780cfb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.686711] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1282.686913] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1282.687601] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef17de95-479a-4960-b98e-235099d1ca67 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.692367] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1282.692367] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5295d1f0-b855-8d8a-9f59-b02c5f02c1b7" [ 1282.692367] env[61986]: _type = "Task" [ 1282.692367] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.698922] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5295d1f0-b855-8d8a-9f59-b02c5f02c1b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.202569] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]5295d1f0-b855-8d8a-9f59-b02c5f02c1b7, 'name': SearchDatastore_Task, 'duration_secs': 0.007975} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.203358] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4519694-7de5-47cd-a7e4-31c89fb2d2cf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.208062] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1283.208062] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520633b4-a1da-d6a9-6a95-3c39ab794775" [ 1283.208062] env[61986]: _type = "Task" [ 1283.208062] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.215609] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520633b4-a1da-d6a9-6a95-3c39ab794775, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.719658] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]520633b4-a1da-d6a9-6a95-3c39ab794775, 'name': SearchDatastore_Task, 'duration_secs': 0.008771} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.719927] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.720215] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec/6ef51778-7ac7-43bc-b6ec-d1e40d0329ec.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1283.720477] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f49b4c0-487d-44a3-8a78-7f22e81e73f3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.726825] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1283.726825] env[61986]: value = "task-1160712" [ 1283.726825] env[61986]: _type = "Task" [ 1283.726825] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.733822] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.266361] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447132} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.266361] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec/6ef51778-7ac7-43bc-b6ec-d1e40d0329ec.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1284.266361] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1284.266361] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da748654-7bc4-4f0d-85d6-00386962a4ed {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.266361] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1284.266361] env[61986]: value = "task-1160713" [ 1284.266361] env[61986]: _type = "Task" [ 1284.266361] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.266361] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160713, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.754545] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160713, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058979} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.754880] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1284.755665] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ff1672-0e5a-4c48-a294-d6d1c811c9b8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.776925] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec/6ef51778-7ac7-43bc-b6ec-d1e40d0329ec.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1284.777171] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b6cb96c-f539-4583-96f8-39449a5ab45e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.796055] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1284.796055] env[61986]: value = "task-1160714" [ 1284.796055] env[61986]: _type = "Task" [ 1284.796055] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.803279] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160714, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.306576] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160714, 'name': ReconfigVM_Task, 'duration_secs': 0.263986} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.306966] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Reconfigured VM instance instance-00000075 to attach disk [datastore2] 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec/6ef51778-7ac7-43bc-b6ec-d1e40d0329ec.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1285.307513] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0601ebf1-2ccb-40c6-a10e-d51a580860c0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.314787] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1285.314787] env[61986]: value = "task-1160715" [ 1285.314787] env[61986]: _type = "Task" [ 1285.314787] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.322363] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160715, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.825013] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160715, 'name': Rename_Task, 'duration_secs': 0.13139} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.825300] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1285.825542] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fdccaadf-b1bc-4ec4-9104-d1be00f948e9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.831533] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1285.831533] env[61986]: value = "task-1160716" [ 1285.831533] env[61986]: _type = "Task" [ 1285.831533] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.838493] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.341531] env[61986]: DEBUG oslo_vmware.api [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160716, 'name': PowerOnVM_Task, 'duration_secs': 0.431349} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.341991] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1286.342062] env[61986]: INFO nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Took 7.80 seconds to spawn the instance on the hypervisor. [ 1286.342217] env[61986]: DEBUG nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1286.342995] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7e51a7-efbd-4e54-ba65-3d092f09ef11 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.863219] env[61986]: INFO nova.compute.manager [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Took 12.96 seconds to build instance. [ 1286.875506] env[61986]: DEBUG nova.compute.manager [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Received event network-changed-3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1286.875506] env[61986]: DEBUG nova.compute.manager [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Refreshing instance network info cache due to event network-changed-3371718c-6628-4377-8a53-64b3ac1baf38. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1286.875683] env[61986]: DEBUG oslo_concurrency.lockutils [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] Acquiring lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.875914] env[61986]: DEBUG oslo_concurrency.lockutils [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] Acquired lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.876179] env[61986]: DEBUG nova.network.neutron [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Refreshing network info cache for port 3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1287.365697] env[61986]: DEBUG oslo_concurrency.lockutils [None req-6b2e72d8-d970-452e-9aa3-8b709a54f416 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.477s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1287.577843] env[61986]: DEBUG nova.network.neutron [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Updated VIF entry in instance network info cache for port 3371718c-6628-4377-8a53-64b3ac1baf38. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1287.578354] env[61986]: DEBUG nova.network.neutron [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Updating instance_info_cache with network_info: [{"id": "3371718c-6628-4377-8a53-64b3ac1baf38", "address": "fa:16:3e:d7:4a:dd", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3371718c-66", "ovs_interfaceid": "3371718c-6628-4377-8a53-64b3ac1baf38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.081527] env[61986]: DEBUG oslo_concurrency.lockutils [req-cfe4204b-d4bc-4781-ab05-090eb384f39c req-d666b3c8-9fea-46d4-ad69-f04cc11820b3 service nova] Releasing lock "refresh_cache-6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1308.186272] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1308.186585] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1308.688961] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1309.210375] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.210662] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1309.212123] env[61986]: INFO nova.compute.claims [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1310.241244] env[61986]: DEBUG nova.scheduler.client.report [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Refreshing inventories for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1310.255121] env[61986]: DEBUG nova.scheduler.client.report [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Updating ProviderTree inventory for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1310.255357] env[61986]: DEBUG nova.compute.provider_tree [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Updating inventory in ProviderTree for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1310.265814] env[61986]: DEBUG nova.scheduler.client.report [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Refreshing aggregate associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, aggregates: None {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1310.282752] env[61986]: DEBUG nova.scheduler.client.report [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Refreshing trait associations for resource provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61986) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1310.323552] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4561ac6d-4784-4cc8-b3cf-8fd4e04c9a7c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.331094] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b813c99e-e7a7-4b31-b792-7cd71e3fe706 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.360698] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c2d3e5-ead9-4407-9933-3c6983d0189e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.367397] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e61c14-1220-4fcc-8eb9-73b34e5936df {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.379609] env[61986]: DEBUG nova.compute.provider_tree [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.882228] env[61986]: DEBUG nova.scheduler.client.report [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1311.387136] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.176s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.387592] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1311.893118] env[61986]: DEBUG nova.compute.utils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1311.894532] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1311.894699] env[61986]: DEBUG nova.network.neutron [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1311.942530] env[61986]: DEBUG nova.policy [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c804063142764cac8244fd3d6bd71e16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '838620f936dc4489be8b99ef87bf37ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1312.202898] env[61986]: DEBUG nova.network.neutron [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Successfully created port: 3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1312.397875] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1312.904836] env[61986]: INFO nova.virt.block_device [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Booting with volume 9f49d9a4-b510-4a00-a931-769a84da844a at /dev/sda [ 1312.941012] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1be2e8ba-1289-48e1-8ee4-9a34a6d5e4d6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.950079] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55fdbea-4df1-48b5-89e6-c7004dc3ff2f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.972632] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7600acf6-1ed2-4aaa-995e-fc8f7842cefa {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.979971] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e9074f-a6fa-4034-b77d-2bc3162f8b84 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.003049] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c67acdd-3a8d-466f-84d8-1c55d91e92e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.008562] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e9437a-748e-4f0a-8bf7-22dc8b6f7cce {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.020812] env[61986]: DEBUG nova.virt.block_device [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating existing volume attachment record: cb9abf3d-fc50-46b1-9cfc-a2314d2348b6 {{(pid=61986) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1313.577187] env[61986]: DEBUG nova.compute.manager [req-f492b6e2-2196-409b-8d43-1f89e0ba0cfc req-f0560d7f-9885-4dda-882b-ecf8987c25da service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Received event network-vif-plugged-3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1313.577476] env[61986]: DEBUG oslo_concurrency.lockutils [req-f492b6e2-2196-409b-8d43-1f89e0ba0cfc req-f0560d7f-9885-4dda-882b-ecf8987c25da service nova] Acquiring lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.577669] env[61986]: DEBUG oslo_concurrency.lockutils [req-f492b6e2-2196-409b-8d43-1f89e0ba0cfc req-f0560d7f-9885-4dda-882b-ecf8987c25da service nova] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.577799] env[61986]: DEBUG oslo_concurrency.lockutils [req-f492b6e2-2196-409b-8d43-1f89e0ba0cfc req-f0560d7f-9885-4dda-882b-ecf8987c25da service nova] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.577961] env[61986]: DEBUG nova.compute.manager [req-f492b6e2-2196-409b-8d43-1f89e0ba0cfc req-f0560d7f-9885-4dda-882b-ecf8987c25da service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] No waiting events found dispatching network-vif-plugged-3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1313.578115] env[61986]: WARNING nova.compute.manager [req-f492b6e2-2196-409b-8d43-1f89e0ba0cfc req-f0560d7f-9885-4dda-882b-ecf8987c25da service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Received unexpected event network-vif-plugged-3ab4cff8-d10f-4916-8f27-a063507d850a for instance with vm_state building and task_state block_device_mapping. [ 1313.690298] env[61986]: DEBUG nova.network.neutron [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Successfully updated port: 3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1314.192138] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1314.192321] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1314.192522] env[61986]: DEBUG nova.network.neutron [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1314.722860] env[61986]: DEBUG nova.network.neutron [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1314.842343] env[61986]: DEBUG nova.network.neutron [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [{"id": "3ab4cff8-d10f-4916-8f27-a063507d850a", "address": "fa:16:3e:f4:a7:52", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab4cff8-d1", "ovs_interfaceid": "3ab4cff8-d10f-4916-8f27-a063507d850a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1315.104194] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1315.104810] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1315.105104] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1315.105322] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1315.105557] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1315.105722] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1315.105881] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1315.106106] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1315.106274] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1315.106447] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1315.106615] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1315.106796] env[61986]: DEBUG nova.virt.hardware [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1315.107697] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38579a96-d106-4004-be2b-712f9f4325e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.115994] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f9b7f5-08dd-4360-b14d-aa91d62ad187 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.345509] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1315.345852] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance network_info: |[{"id": "3ab4cff8-d10f-4916-8f27-a063507d850a", "address": "fa:16:3e:f4:a7:52", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab4cff8-d1", "ovs_interfaceid": "3ab4cff8-d10f-4916-8f27-a063507d850a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1315.346363] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:a7:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ab4cff8-d10f-4916-8f27-a063507d850a', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1315.353904] env[61986]: DEBUG oslo.service.loopingcall [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1315.354131] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1315.354362] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1005c87a-d6f4-4566-b0e5-ba1af5276b0f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.373748] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1315.373748] env[61986]: value = "task-1160725" [ 1315.373748] env[61986]: _type = "Task" [ 1315.373748] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.380871] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160725, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.602560] env[61986]: DEBUG nova.compute.manager [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Received event network-changed-3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1315.602785] env[61986]: DEBUG nova.compute.manager [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Refreshing instance network info cache due to event network-changed-3ab4cff8-d10f-4916-8f27-a063507d850a. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1315.602999] env[61986]: DEBUG oslo_concurrency.lockutils [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] Acquiring lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1315.603162] env[61986]: DEBUG oslo_concurrency.lockutils [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] Acquired lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1315.603355] env[61986]: DEBUG nova.network.neutron [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Refreshing network info cache for port 3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1315.883735] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160725, 'name': CreateVM_Task, 'duration_secs': 0.293318} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.884112] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1315.884608] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252506', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'name': 'volume-9f49d9a4-b510-4a00-a931-769a84da844a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad', 'attached_at': '', 'detached_at': '', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'serial': '9f49d9a4-b510-4a00-a931-769a84da844a'}, 'guest_format': None, 'attachment_id': 'cb9abf3d-fc50-46b1-9cfc-a2314d2348b6', 'boot_index': 0, 'disk_bus': None, 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61986) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1315.884859] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Root volume attach. Driver type: vmdk {{(pid=61986) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1315.885633] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c08d1d-d7da-4b2d-badb-c552a67f85f5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.892769] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e13e5d-bf20-4501-90cd-a8ca3040b52a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.898407] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf818f9d-8156-4d9a-8b82-79233758d74e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.903793] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-6f289d7a-78ab-4dc0-947f-2af305cbe8e6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.910386] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1315.910386] env[61986]: value = "task-1160726" [ 1315.910386] env[61986]: _type = "Task" [ 1315.910386] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.917589] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.320329] env[61986]: DEBUG nova.network.neutron [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updated VIF entry in instance network info cache for port 3ab4cff8-d10f-4916-8f27-a063507d850a. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1316.320544] env[61986]: DEBUG nova.network.neutron [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [{"id": "3ab4cff8-d10f-4916-8f27-a063507d850a", "address": "fa:16:3e:f4:a7:52", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab4cff8-d1", "ovs_interfaceid": "3ab4cff8-d10f-4916-8f27-a063507d850a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.421647] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task} progress is 43%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.824084] env[61986]: DEBUG oslo_concurrency.lockutils [req-10f5c26e-d326-4e0d-b7f7-ea6d6b361631 req-92390a86-bcb8-4fb9-9bf0-916d04334252 service nova] Releasing lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1316.921091] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task} progress is 60%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.422405] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task} progress is 75%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.923981] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task} progress is 92%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.423418] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task} progress is 98%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.925109] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160726, 'name': RelocateVM_Task, 'duration_secs': 2.787735} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.925495] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Volume attach. Driver type: vmdk {{(pid=61986) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1318.925709] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252506', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'name': 'volume-9f49d9a4-b510-4a00-a931-769a84da844a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad', 'attached_at': '', 'detached_at': '', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'serial': '9f49d9a4-b510-4a00-a931-769a84da844a'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1318.926477] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6560d6be-9aa1-472f-8e43-569dcad7af0a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.941473] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12b82a9-4e0d-4daf-aebe-98c98fba2a0b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.963250] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] volume-9f49d9a4-b510-4a00-a931-769a84da844a/volume-9f49d9a4-b510-4a00-a931-769a84da844a.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1318.963486] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-beab4069-3ced-468e-93f1-9334597f9516 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.981776] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1318.981776] env[61986]: value = "task-1160727" [ 1318.981776] env[61986]: _type = "Task" [ 1318.981776] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.988934] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160727, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.491581] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.992062] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160727, 'name': ReconfigVM_Task, 'duration_secs': 0.755721} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.992434] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Reconfigured VM instance instance-00000076 to attach disk [datastore1] volume-9f49d9a4-b510-4a00-a931-769a84da844a/volume-9f49d9a4-b510-4a00-a931-769a84da844a.vmdk or device None with type thin {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1319.996989] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9821ca0c-8b0b-4483-80c5-8dd628b3bf9f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.011868] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1320.011868] env[61986]: value = "task-1160728" [ 1320.011868] env[61986]: _type = "Task" [ 1320.011868] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.019387] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160728, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.521808] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160728, 'name': ReconfigVM_Task, 'duration_secs': 0.10696} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.522145] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252506', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'name': 'volume-9f49d9a4-b510-4a00-a931-769a84da844a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad', 'attached_at': '', 'detached_at': '', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'serial': '9f49d9a4-b510-4a00-a931-769a84da844a'} {{(pid=61986) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1320.522690] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85297b29-0aed-4fe4-b639-0ba89b861e08 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.528896] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1320.528896] env[61986]: value = "task-1160729" [ 1320.528896] env[61986]: _type = "Task" [ 1320.528896] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.542094] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160729, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.038736] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160729, 'name': Rename_Task, 'duration_secs': 0.118491} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.039124] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1321.039386] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28174d5f-48b6-4e76-a19a-87358cf44740 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.045512] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1321.045512] env[61986]: value = "task-1160730" [ 1321.045512] env[61986]: _type = "Task" [ 1321.045512] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.052582] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160730, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.555758] env[61986]: DEBUG oslo_vmware.api [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160730, 'name': PowerOnVM_Task, 'duration_secs': 0.447935} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.556047] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1321.556249] env[61986]: INFO nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 6.45 seconds to spawn the instance on the hypervisor. [ 1321.556431] env[61986]: DEBUG nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1321.557191] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0662620-db60-447c-bf02-e3403bda7e10 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.077108] env[61986]: INFO nova.compute.manager [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 12.88 seconds to build instance. [ 1322.475336] env[61986]: DEBUG nova.compute.manager [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Received event network-changed-5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1322.475525] env[61986]: DEBUG nova.compute.manager [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Refreshing instance network info cache due to event network-changed-5de45ae0-ff50-4fae-8941-3bd85428a59e. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1322.475745] env[61986]: DEBUG oslo_concurrency.lockutils [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] Acquiring lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1322.475893] env[61986]: DEBUG oslo_concurrency.lockutils [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] Acquired lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1322.476195] env[61986]: DEBUG nova.network.neutron [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Refreshing network info cache for port 5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1322.579847] env[61986]: DEBUG oslo_concurrency.lockutils [None req-e39ebd57-8abd-4c17-8f1f-c35a347f63ae tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.393s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.188946] env[61986]: DEBUG nova.network.neutron [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updated VIF entry in instance network info cache for port 5de45ae0-ff50-4fae-8941-3bd85428a59e. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1323.189515] env[61986]: DEBUG nova.network.neutron [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [{"id": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "address": "fa:16:3e:22:d2:28", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5de45ae0-ff", "ovs_interfaceid": "5de45ae0-ff50-4fae-8941-3bd85428a59e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1323.692924] env[61986]: DEBUG oslo_concurrency.lockutils [req-9ed65185-e487-41b3-8022-5e4d99a9d199 req-ef616e51-d075-40a7-9c80-0f3768d1de40 service nova] Releasing lock "refresh_cache-a6094cc4-7d17-4858-abae-972425241c8f" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1323.879082] env[61986]: DEBUG nova.compute.manager [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Stashing vm_state: active {{(pid=61986) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1324.402086] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1324.402362] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1324.499943] env[61986]: DEBUG nova.compute.manager [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Received event network-changed-3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1324.500181] env[61986]: DEBUG nova.compute.manager [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Refreshing instance network info cache due to event network-changed-3ab4cff8-d10f-4916-8f27-a063507d850a. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1324.500399] env[61986]: DEBUG oslo_concurrency.lockutils [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] Acquiring lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1324.500598] env[61986]: DEBUG oslo_concurrency.lockutils [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] Acquired lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.500725] env[61986]: DEBUG nova.network.neutron [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Refreshing network info cache for port 3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1324.907626] env[61986]: INFO nova.compute.claims [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1325.215401] env[61986]: DEBUG nova.network.neutron [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updated VIF entry in instance network info cache for port 3ab4cff8-d10f-4916-8f27-a063507d850a. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1325.215769] env[61986]: DEBUG nova.network.neutron [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [{"id": "3ab4cff8-d10f-4916-8f27-a063507d850a", "address": "fa:16:3e:f4:a7:52", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab4cff8-d1", "ovs_interfaceid": "3ab4cff8-d10f-4916-8f27-a063507d850a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1325.414343] env[61986]: INFO nova.compute.resource_tracker [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating resource usage from migration 6a7775a8-0fa1-4166-9965-faadb4932037 [ 1325.470194] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eaa0dd1-3420-4fc7-9292-199d0e3f4241 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.477656] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1789941d-3dd6-4784-ae77-167219a68361 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.508332] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1b5166-3bb5-4fcc-a242-7b377ec27c31 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.515247] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4945750a-b586-49ec-8baf-d480eb823b7e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.527817] env[61986]: DEBUG nova.compute.provider_tree [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1325.718769] env[61986]: DEBUG oslo_concurrency.lockutils [req-305e6a26-9676-4ccd-9dce-b6d12ca9a879 req-a06d4a72-a744-4bce-a5b6-259d82395db2 service nova] Releasing lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.030996] env[61986]: DEBUG nova.scheduler.client.report [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1326.536380] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.134s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.536763] env[61986]: INFO nova.compute.manager [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Migrating [ 1327.050985] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1327.051213] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquired lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1327.051420] env[61986]: DEBUG nova.network.neutron [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1327.754871] env[61986]: DEBUG nova.network.neutron [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [{"id": "3ab4cff8-d10f-4916-8f27-a063507d850a", "address": "fa:16:3e:f4:a7:52", "network": {"id": "fa5ac770-eddf-47d9-9a5c-e8131367c908", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1291662626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "838620f936dc4489be8b99ef87bf37ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ab4cff8-d1", "ovs_interfaceid": "3ab4cff8-d10f-4916-8f27-a063507d850a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1328.258122] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Releasing lock "refresh_cache-ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1329.773289] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff68b912-567a-4d8f-9f75-46ae2e8de396 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.792787] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad' progress to 0 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1330.214594] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.298863] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1330.299136] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-725220f1-c712-4556-99e7-cc4025885940 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.307240] env[61986]: DEBUG oslo_vmware.api [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1330.307240] env[61986]: value = "task-1160731" [ 1330.307240] env[61986]: _type = "Task" [ 1330.307240] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.317945] env[61986]: DEBUG oslo_vmware.api [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.817039] env[61986]: DEBUG oslo_vmware.api [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160731, 'name': PowerOffVM_Task, 'duration_secs': 0.223053} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.817406] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1330.817511] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad' progress to 17 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1331.213567] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.213823] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.213971] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1331.214134] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1331.324029] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1331.324354] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1331.324520] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1331.324710] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1331.324860] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1331.325031] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1331.325309] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1331.325497] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1331.325670] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1331.325836] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1331.326022] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1331.331970] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4b4037e-06b2-4b3b-bfa7-ceb298ab011f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.348517] env[61986]: DEBUG oslo_vmware.api [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1331.348517] env[61986]: value = "task-1160732" [ 1331.348517] env[61986]: _type = "Task" [ 1331.348517] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.356478] env[61986]: DEBUG oslo_vmware.api [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160732, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.717780] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.718053] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.718258] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1331.718421] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1331.719340] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07113ad4-c7dc-4cff-b173-93fb03ab78ee {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.727319] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cc03cb-d0a0-4b18-8df9-113b447ac32f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.740873] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d062c71-8f00-4c87-a385-9075b91316e5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.747130] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0280d28-6110-4efa-99e7-d43fa72e94d1 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.775879] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181162MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1331.776046] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.776218] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.858191] env[61986]: DEBUG oslo_vmware.api [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160732, 'name': ReconfigVM_Task, 'duration_secs': 0.180055} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.858603] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad' progress to 33 {{(pid=61986) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1332.364257] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1332.364532] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1332.364670] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1332.364853] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1332.365012] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1332.365166] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1332.365365] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1332.365521] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1332.365682] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1332.365843] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1332.366022] env[61986]: DEBUG nova.virt.hardware [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1332.784044] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Applying migration context for instance ad1da2dc-e0f3-4bb9-8df7-82039ec53aad as it has an incoming, in-progress migration 6a7775a8-0fa1-4166-9965-faadb4932037. Migration status is migrating {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1332.784161] env[61986]: INFO nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating resource usage from migration 6a7775a8-0fa1-4166-9965-faadb4932037 [ 1332.800398] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance a6094cc4-7d17-4858-abae-972425241c8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.800540] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.800665] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Migration 6a7775a8-0fa1-4166-9965-faadb4932037 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1332.800823] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Instance ad1da2dc-e0f3-4bb9-8df7-82039ec53aad actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61986) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1332.801014] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1332.801167] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1332.852629] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559f4f07-2a67-46e6-b013-a5f7456064a7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.859748] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da667f8c-2ccc-46ab-a7e3-380976510708 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.888544] env[61986]: ERROR nova.compute.manager [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Traceback (most recent call last): [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] yield [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] disk_info = self.driver.migrate_disk_and_power_off( [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 263, in migrate_disk_and_power_off [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] disk_key = device.key [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] AttributeError: 'NoneType' object has no attribute 'key' [ 1332.888544] env[61986]: ERROR nova.compute.manager [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] [ 1332.892419] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f49ef4-b5c1-4739-91c5-667429209645 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.899758] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25e7c8f-6086-4c41-a7f0-9054dd984dbf {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.912214] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1333.408012] env[61986]: INFO nova.compute.manager [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Swapping old allocation on dict_keys(['2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d']) held by migration 6a7775a8-0fa1-4166-9965-faadb4932037 for instance [ 1333.414701] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1333.428778] env[61986]: DEBUG nova.scheduler.client.report [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Overwriting current allocation {'allocations': {'2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 159}}, 'project_id': '838620f936dc4489be8b99ef87bf37ce', 'user_id': 'c804063142764cac8244fd3d6bd71e16', 'consumer_generation': 1} on consumer ad1da2dc-e0f3-4bb9-8df7-82039ec53aad {{(pid=61986) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1333.919578] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1333.919814] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.144s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1334.591734] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.591982] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.592201] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.592387] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.592560] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1334.595024] env[61986]: INFO nova.compute.manager [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Terminating instance [ 1334.597057] env[61986]: DEBUG nova.compute.manager [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1334.597280] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1334.597517] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2debc7e-01cc-4c50-9525-8f8cb46036d9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.604319] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1334.604319] env[61986]: value = "task-1160733" [ 1334.604319] env[61986]: _type = "Task" [ 1334.604319] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.612034] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.982077] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.982474] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1335.034207] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63964fd-7b7c-4eae-bc6c-81f5ae1d2480 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.041538] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c4ebe5-3559-4d68-9929-40e89854c3e2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.071187] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0a2343-ff59-4ed5-bf3b-c6e9a11fb35d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.077529] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145d4c2e-0c51-4d4d-9d3b-b37c1c885bc8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.089818] env[61986]: DEBUG nova.compute.provider_tree [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1335.112794] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] VM already powered off {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1335.113039] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Volume detach. Driver type: vmdk {{(pid=61986) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1335.113241] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252506', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'name': 'volume-9f49d9a4-b510-4a00-a931-769a84da844a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad', 'attached_at': '', 'detached_at': '', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'serial': '9f49d9a4-b510-4a00-a931-769a84da844a'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1335.114167] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09254352-8ad1-44c3-8450-b495b6d14437 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.131283] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd22483-e2ea-44c9-9600-29ffe7f7f925 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.137578] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7e3d97-9bf9-4181-ab73-1a9a95e35d5d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.153797] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c187b6e-37b2-472d-bc74-be45cf1d39f4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.167732] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] The volume has not been displaced from its original location: [datastore1] volume-9f49d9a4-b510-4a00-a931-769a84da844a/volume-9f49d9a4-b510-4a00-a931-769a84da844a.vmdk. No consolidation needed. {{(pid=61986) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1335.172807] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Reconfiguring VM instance instance-00000076 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1335.173089] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5182daf0-8e20-4ea8-b5f1-bffc09b8a34c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.189130] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1335.189130] env[61986]: value = "task-1160734" [ 1335.189130] env[61986]: _type = "Task" [ 1335.189130] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.195933] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160734, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.593273] env[61986]: DEBUG nova.scheduler.client.report [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1335.700068] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160734, 'name': ReconfigVM_Task, 'duration_secs': 0.143935} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.700068] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Reconfigured VM instance instance-00000076 to detach disk 2000 {{(pid=61986) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1335.704623] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d3c0ddb-b052-4a15-92b9-ef3aedb92b2c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.718847] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1335.718847] env[61986]: value = "task-1160735" [ 1335.718847] env[61986]: _type = "Task" [ 1335.718847] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.726228] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.907972] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1335.921461] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1335.921461] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1335.921461] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1336.098914] env[61986]: DEBUG oslo_concurrency.lockutils [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.116s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.099397] env[61986]: INFO nova.compute.manager [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Successfully reverted task state from resize_migrating on failure for instance. [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server [None req-f1572e1a-5719-4b9c-9ab0-36ab2cdc9558 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server raise self.value [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server raise self.value [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server raise self.value [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server raise self.value [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 263, in migrate_disk_and_power_off [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1336.107879] env[61986]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1336.109928] env[61986]: ERROR oslo_messaging.rpc.server [ 1336.229166] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160735, 'name': ReconfigVM_Task, 'duration_secs': 0.110953} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.229416] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-252506', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'name': 'volume-9f49d9a4-b510-4a00-a931-769a84da844a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad1da2dc-e0f3-4bb9-8df7-82039ec53aad', 'attached_at': '', 'detached_at': '', 'volume_id': '9f49d9a4-b510-4a00-a931-769a84da844a', 'serial': '9f49d9a4-b510-4a00-a931-769a84da844a'} {{(pid=61986) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1336.229679] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1336.230446] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a9559b-1a52-4150-8f33-bca988ca9d73 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.236846] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1336.237092] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5187b418-976e-440c-942f-d3670a012d53 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.295654] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1336.295889] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Deleting contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1336.296050] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore1] ad1da2dc-e0f3-4bb9-8df7-82039ec53aad {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1336.296321] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4de4cd5c-4fcd-45dd-8de2-16a6cee91f3b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.302525] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1336.302525] env[61986]: value = "task-1160737" [ 1336.302525] env[61986]: _type = "Task" [ 1336.302525] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.309975] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.812461] env[61986]: DEBUG oslo_vmware.api [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.075832} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.812663] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1336.812853] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Deleted contents of the VM from datastore datastore1 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1336.813065] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1336.813250] env[61986]: INFO nova.compute.manager [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1336.813488] env[61986]: DEBUG oslo.service.loopingcall [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1336.813673] env[61986]: DEBUG nova.compute.manager [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1336.813778] env[61986]: DEBUG nova.network.neutron [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1337.194016] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.194454] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1337.194763] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.195053] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1337.195337] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.197862] env[61986]: INFO nova.compute.manager [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Terminating instance [ 1337.199744] env[61986]: DEBUG nova.compute.manager [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1337.200020] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1337.201270] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f132dc5-82bb-4131-bad6-a6b7735c7ebe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.208888] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1337.209130] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aef03831-df2a-48c4-aef6-1c4b5129fb99 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.214402] env[61986]: DEBUG oslo_vmware.api [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1337.214402] env[61986]: value = "task-1160738" [ 1337.214402] env[61986]: _type = "Task" [ 1337.214402] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.223883] env[61986]: DEBUG oslo_vmware.api [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.279952] env[61986]: DEBUG nova.compute.manager [req-6d95f742-a7e3-4ff2-b0c2-f0be41d2792f req-c9c4e9a8-bb30-44a0-b4e1-0f1c12cbb081 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Received event network-vif-deleted-3ab4cff8-d10f-4916-8f27-a063507d850a {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1337.280238] env[61986]: INFO nova.compute.manager [req-6d95f742-a7e3-4ff2-b0c2-f0be41d2792f req-c9c4e9a8-bb30-44a0-b4e1-0f1c12cbb081 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Neutron deleted interface 3ab4cff8-d10f-4916-8f27-a063507d850a; detaching it from the instance and deleting it from the info cache [ 1337.280426] env[61986]: DEBUG nova.network.neutron [req-6d95f742-a7e3-4ff2-b0c2-f0be41d2792f req-c9c4e9a8-bb30-44a0-b4e1-0f1c12cbb081 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.429910] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Didn't find any instances for network info cache update. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1337.430241] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.430511] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.430751] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.724732] env[61986]: DEBUG oslo_vmware.api [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160738, 'name': PowerOffVM_Task, 'duration_secs': 0.168009} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.725060] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1337.725206] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1337.725457] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-304766c3-c6b0-43fd-a2d6-cb31f31aba2b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.747770] env[61986]: DEBUG nova.network.neutron [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.783395] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2cae6e7-8265-4cb2-be80-14725a173002 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.794047] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43dbcd82-4e80-47ff-96c7-43e0f422ea19 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.808534] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1337.808771] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1337.808958] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore2] 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1337.809849] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f7e2b11-0de3-4813-89f9-bef5da0c6bb0 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.817229] env[61986]: DEBUG oslo_vmware.api [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1337.817229] env[61986]: value = "task-1160740" [ 1337.817229] env[61986]: _type = "Task" [ 1337.817229] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.825303] env[61986]: DEBUG nova.compute.manager [req-6d95f742-a7e3-4ff2-b0c2-f0be41d2792f req-c9c4e9a8-bb30-44a0-b4e1-0f1c12cbb081 service nova] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Detach interface failed, port_id=3ab4cff8-d10f-4916-8f27-a063507d850a, reason: Instance ad1da2dc-e0f3-4bb9-8df7-82039ec53aad could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1337.830720] env[61986]: DEBUG oslo_vmware.api [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.251081] env[61986]: INFO nova.compute.manager [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 1.44 seconds to deallocate network for instance. [ 1338.327678] env[61986]: DEBUG oslo_vmware.api [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133551} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.327945] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1338.328161] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1338.328370] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1338.328606] env[61986]: INFO nova.compute.manager [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1338.328865] env[61986]: DEBUG oslo.service.loopingcall [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1338.329082] env[61986]: DEBUG nova.compute.manager [-] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1338.329182] env[61986]: DEBUG nova.network.neutron [-] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1338.805291] env[61986]: INFO nova.compute.manager [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 0.55 seconds to detach 1 volumes for instance. [ 1338.807635] env[61986]: DEBUG nova.compute.manager [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Deleting volume: 9f49d9a4-b510-4a00-a931-769a84da844a {{(pid=61986) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1339.053940] env[61986]: DEBUG nova.network.neutron [-] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1339.303548] env[61986]: DEBUG nova.compute.manager [req-fc5814c4-95b0-4897-888a-a74983cdbf04 req-211095a5-e2e9-44bc-8f08-2887e46cdf10 service nova] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Received event network-vif-deleted-3371718c-6628-4377-8a53-64b3ac1baf38 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1339.348732] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1339.348997] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.349255] env[61986]: DEBUG nova.objects.instance [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'resources' on Instance uuid ad1da2dc-e0f3-4bb9-8df7-82039ec53aad {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1339.556254] env[61986]: INFO nova.compute.manager [-] [instance: 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec] Took 1.23 seconds to deallocate network for instance. [ 1339.903355] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14e54e2-a9e7-4a68-bc44-775c4c14c3c7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.910795] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e24ffeb-6ed5-45f7-ad79-4c30ccd24d8f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.940041] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949cd004-c2c5-4a20-9e25-6ead5e429e03 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.947082] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b61e70b-7801-4c0f-9f93-b6d66597a400 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.959741] env[61986]: DEBUG nova.compute.provider_tree [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.062740] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.462824] env[61986]: DEBUG nova.scheduler.client.report [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1340.968515] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.971025] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.908s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.971120] env[61986]: DEBUG nova.objects.instance [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'resources' on Instance uuid 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1340.985697] env[61986]: INFO nova.scheduler.client.report [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted allocations for instance ad1da2dc-e0f3-4bb9-8df7-82039ec53aad [ 1341.494058] env[61986]: DEBUG oslo_concurrency.lockutils [None req-a1830d4b-038a-4cad-bf49-b22555dacd6d tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.902s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.494917] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.587s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.495169] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.495378] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.495548] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.498316] env[61986]: INFO nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Terminating instance [ 1341.500083] env[61986]: DEBUG nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1341.500366] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86af94f3-4ce8-4568-b516-ff29f2e55266 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.508848] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864d2cef-a99b-4025-a2d0-78fd7ac642f2 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.520704] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2823d35f-d9bc-4986-9cc7-73fcbe23ba9a {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.527843] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b11a4cc-1a2a-40b0-a388-3ce274e9d987 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.535017] env[61986]: WARNING nova.virt.vmwareapi.driver [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance ad1da2dc-e0f3-4bb9-8df7-82039ec53aad could not be found. [ 1341.535240] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1341.535774] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3065135a-c112-4581-83ec-99c94082f834 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.562065] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc89d9a-c99a-4300-8ea0-5c0e3454134f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.566911] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefec04f-3bff-4c65-a62d-57331d78be33 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.580852] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2c6d40-de0d-4bb5-9847-aa27908a9eb8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.589941] env[61986]: WARNING nova.virt.vmwareapi.vmops [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad1da2dc-e0f3-4bb9-8df7-82039ec53aad could not be found. [ 1341.590143] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1341.590322] env[61986]: INFO nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 0.09 seconds to destroy the instance on the hypervisor. [ 1341.590552] env[61986]: DEBUG oslo.service.loopingcall [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1341.597739] env[61986]: DEBUG nova.compute.manager [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1341.597835] env[61986]: DEBUG nova.network.neutron [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1341.599492] env[61986]: DEBUG nova.compute.provider_tree [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1342.102309] env[61986]: DEBUG nova.scheduler.client.report [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1342.115395] env[61986]: DEBUG nova.network.neutron [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1342.607220] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.636s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.617718] env[61986]: INFO nova.compute.manager [-] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 1.02 seconds to deallocate network for instance. [ 1342.628617] env[61986]: INFO nova.scheduler.client.report [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted allocations for instance 6ef51778-7ac7-43bc-b6ec-d1e40d0329ec [ 1342.632698] env[61986]: WARNING nova.volume.cinder [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Attachment cb9abf3d-fc50-46b1-9cfc-a2314d2348b6 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = cb9abf3d-fc50-46b1-9cfc-a2314d2348b6. (HTTP 404) (Request-ID: req-47210b13-c25d-44b0-bf20-23b6510aa921) [ 1342.633479] env[61986]: INFO nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Took 0.01 seconds to detach 1 volumes for instance. [ 1342.636712] env[61986]: DEBUG nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Deleting volume: 9f49d9a4-b510-4a00-a931-769a84da844a {{(pid=61986) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1342.651985] env[61986]: WARNING nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Failed to delete volume: 9f49d9a4-b510-4a00-a931-769a84da844a due to Volume 9f49d9a4-b510-4a00-a931-769a84da844a could not be found.: nova.exception.VolumeNotFound: Volume 9f49d9a4-b510-4a00-a931-769a84da844a could not be found. [ 1343.137940] env[61986]: DEBUG oslo_concurrency.lockutils [None req-84d6eb1f-8d6c-4440-9faf-c58ed5f1a783 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "6ef51778-7ac7-43bc-b6ec-d1e40d0329ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.943s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1343.157597] env[61986]: INFO nova.compute.manager [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: ad1da2dc-e0f3-4bb9-8df7-82039ec53aad] Instance disappeared during terminate [ 1343.157741] env[61986]: DEBUG oslo_concurrency.lockutils [None req-7f02468b-3c3d-47db-8ed4-60ad41ebeefe tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "ad1da2dc-e0f3-4bb9-8df7-82039ec53aad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.663s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.830462] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "a6094cc4-7d17-4858-abae-972425241c8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.830888] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "a6094cc4-7d17-4858-abae-972425241c8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.830935] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "a6094cc4-7d17-4858-abae-972425241c8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.831169] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "a6094cc4-7d17-4858-abae-972425241c8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.831352] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "a6094cc4-7d17-4858-abae-972425241c8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.833530] env[61986]: INFO nova.compute.manager [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Terminating instance [ 1346.835334] env[61986]: DEBUG nova.compute.manager [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1346.835542] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1346.836388] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c26981-3ea4-42c2-97fc-7715012c89fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.844130] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1346.844358] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-238f24e1-9243-4b91-a87e-e65204fe04b5 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.850528] env[61986]: DEBUG oslo_vmware.api [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1346.850528] env[61986]: value = "task-1160743" [ 1346.850528] env[61986]: _type = "Task" [ 1346.850528] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.857727] env[61986]: DEBUG oslo_vmware.api [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.361029] env[61986]: DEBUG oslo_vmware.api [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160743, 'name': PowerOffVM_Task, 'duration_secs': 0.192134} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.361352] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1347.362057] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1347.362057] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4910c789-e5c5-4b5a-b2dc-c0edd7983dcb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.420452] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1347.420617] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1347.420793] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleting the datastore file [datastore2] a6094cc4-7d17-4858-abae-972425241c8f {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1347.421069] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29aacdb7-6150-483c-99f8-0c3a932eb991 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.426953] env[61986]: DEBUG oslo_vmware.api [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for the task: (returnval){ [ 1347.426953] env[61986]: value = "task-1160745" [ 1347.426953] env[61986]: _type = "Task" [ 1347.426953] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.434016] env[61986]: DEBUG oslo_vmware.api [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.938402] env[61986]: DEBUG oslo_vmware.api [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Task: {'id': task-1160745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154846} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.938810] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1347.938858] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1347.939012] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1347.939216] env[61986]: INFO nova.compute.manager [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1347.939465] env[61986]: DEBUG oslo.service.loopingcall [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1347.939660] env[61986]: DEBUG nova.compute.manager [-] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1347.939758] env[61986]: DEBUG nova.network.neutron [-] [instance: a6094cc4-7d17-4858-abae-972425241c8f] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1348.202013] env[61986]: DEBUG nova.compute.manager [req-d03fa14c-5084-4a52-b276-1eec3e27b5a4 req-930da93e-81da-4f8a-b8ac-eb6a38ca2c3e service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Received event network-vif-deleted-5de45ae0-ff50-4fae-8941-3bd85428a59e {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1348.202235] env[61986]: INFO nova.compute.manager [req-d03fa14c-5084-4a52-b276-1eec3e27b5a4 req-930da93e-81da-4f8a-b8ac-eb6a38ca2c3e service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Neutron deleted interface 5de45ae0-ff50-4fae-8941-3bd85428a59e; detaching it from the instance and deleting it from the info cache [ 1348.202412] env[61986]: DEBUG nova.network.neutron [req-d03fa14c-5084-4a52-b276-1eec3e27b5a4 req-930da93e-81da-4f8a-b8ac-eb6a38ca2c3e service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1348.681875] env[61986]: DEBUG nova.network.neutron [-] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1348.704761] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64515aa2-36a6-43ee-925f-c47723656f23 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.714609] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4e32ba-73f3-4f0c-814e-65b5500cc7f6 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.735741] env[61986]: DEBUG nova.compute.manager [req-d03fa14c-5084-4a52-b276-1eec3e27b5a4 req-930da93e-81da-4f8a-b8ac-eb6a38ca2c3e service nova] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Detach interface failed, port_id=5de45ae0-ff50-4fae-8941-3bd85428a59e, reason: Instance a6094cc4-7d17-4858-abae-972425241c8f could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1349.185029] env[61986]: INFO nova.compute.manager [-] [instance: a6094cc4-7d17-4858-abae-972425241c8f] Took 1.25 seconds to deallocate network for instance. [ 1349.691762] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1349.692065] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1349.692326] env[61986]: DEBUG nova.objects.instance [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lazy-loading 'resources' on Instance uuid a6094cc4-7d17-4858-abae-972425241c8f {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1350.226925] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ab2b3a-c454-4ea8-a6a4-4b166092ac44 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.234559] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34732216-8dc6-4182-a254-3c001abd5e40 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.264601] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3b743b-c293-4b4a-9e4d-eeaa4374c21b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.271083] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799ced59-abbb-4961-adef-2861323bb353 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.283481] env[61986]: DEBUG nova.compute.provider_tree [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1350.786147] env[61986]: DEBUG nova.scheduler.client.report [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1351.291948] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1351.311015] env[61986]: INFO nova.scheduler.client.report [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Deleted allocations for instance a6094cc4-7d17-4858-abae-972425241c8f [ 1351.819216] env[61986]: DEBUG oslo_concurrency.lockutils [None req-8d9ab8bf-4a95-4e0d-ac70-987ad9ea4d42 tempest-ServerActionsTestOtherA-1979745198 tempest-ServerActionsTestOtherA-1979745198-project-member] Lock "a6094cc4-7d17-4858-abae-972425241c8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.988s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.533777] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.534102] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.036676] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Starting instance... {{(pid=61986) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1357.562782] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.563126] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.564920] env[61986]: INFO nova.compute.claims [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1358.603800] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88609cc1-9438-4d25-bd3d-e840997e288b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.611646] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d78d9b5-61d3-493f-8ba4-16cdead7a8ab {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.640680] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81244e6f-f748-4a89-a30e-083b8e64d75d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.647725] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a56c5d-000c-457f-8919-2f7cae53b659 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.660139] env[61986]: DEBUG nova.compute.provider_tree [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1359.163593] env[61986]: DEBUG nova.scheduler.client.report [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1359.668353] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1359.668914] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Start building networks asynchronously for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1360.174511] env[61986]: DEBUG nova.compute.utils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Using /dev/sd instead of None {{(pid=61986) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1360.176304] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Allocating IP information in the background. {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1360.176484] env[61986]: DEBUG nova.network.neutron [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] allocate_for_instance() {{(pid=61986) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1360.215501] env[61986]: DEBUG nova.policy [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b87e056b5450424886dcc1eb6163068b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc80e86b089d4c8c92b721d8fa944fb5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61986) authorize /opt/stack/nova/nova/policy.py:203}} [ 1360.453301] env[61986]: DEBUG nova.network.neutron [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Successfully created port: 10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1360.679716] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Start building block device mappings for instance. {{(pid=61986) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1361.688555] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Start spawning the instance on the hypervisor. {{(pid=61986) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1361.715976] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T06:37:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T06:37:12Z,direct_url=,disk_format='vmdk',id=7b73cf0e-555e-4258-a7ba-6cecbc90b6c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='f93eeac37dc4477cb79aae49b3cd744b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T06:37:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1361.716264] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Flavor limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1361.716426] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Image limits 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1361.716619] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Flavor pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1361.716813] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Image pref 0:0:0 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1361.716908] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61986) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1361.717128] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1361.717296] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1361.717463] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Got 1 possible topologies {{(pid=61986) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1361.717625] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1361.717798] env[61986]: DEBUG nova.virt.hardware [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61986) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1361.718781] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1666cc-acf4-4e8d-b222-2557321fab42 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.726896] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5aab6f7-c06d-44b5-a7ea-ca46d831b6fb {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.826504] env[61986]: DEBUG nova.compute.manager [req-88b619e7-491e-445a-aa8a-5c47ba4f1f11 req-5bf29c05-3d75-46a0-8bc8-8c7c614aaa5b service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Received event network-vif-plugged-10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1361.826737] env[61986]: DEBUG oslo_concurrency.lockutils [req-88b619e7-491e-445a-aa8a-5c47ba4f1f11 req-5bf29c05-3d75-46a0-8bc8-8c7c614aaa5b service nova] Acquiring lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1361.826945] env[61986]: DEBUG oslo_concurrency.lockutils [req-88b619e7-491e-445a-aa8a-5c47ba4f1f11 req-5bf29c05-3d75-46a0-8bc8-8c7c614aaa5b service nova] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1361.827142] env[61986]: DEBUG oslo_concurrency.lockutils [req-88b619e7-491e-445a-aa8a-5c47ba4f1f11 req-5bf29c05-3d75-46a0-8bc8-8c7c614aaa5b service nova] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1361.827355] env[61986]: DEBUG nova.compute.manager [req-88b619e7-491e-445a-aa8a-5c47ba4f1f11 req-5bf29c05-3d75-46a0-8bc8-8c7c614aaa5b service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] No waiting events found dispatching network-vif-plugged-10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1361.827484] env[61986]: WARNING nova.compute.manager [req-88b619e7-491e-445a-aa8a-5c47ba4f1f11 req-5bf29c05-3d75-46a0-8bc8-8c7c614aaa5b service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Received unexpected event network-vif-plugged-10d3e612-fc06-4163-9744-de41c9a9a307 for instance with vm_state building and task_state spawning. [ 1361.909045] env[61986]: DEBUG nova.network.neutron [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Successfully updated port: 10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1362.411150] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "refresh_cache-6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1362.411351] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquired lock "refresh_cache-6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1362.411483] env[61986]: DEBUG nova.network.neutron [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Building network info cache for instance {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1362.942530] env[61986]: DEBUG nova.network.neutron [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Instance cache missing network info. {{(pid=61986) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1363.063097] env[61986]: DEBUG nova.network.neutron [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Updating instance_info_cache with network_info: [{"id": "10d3e612-fc06-4163-9744-de41c9a9a307", "address": "fa:16:3e:8b:d3:4e", "network": {"id": "ec789ef1-85c7-4760-b9d7-cffbbdc71557", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-466404986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc80e86b089d4c8c92b721d8fa944fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d177c5b3-a5b1-4c78-854e-7e0dbf341ea1", "external-id": "nsx-vlan-transportzone-54", "segmentation_id": 54, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d3e612-fc", "ovs_interfaceid": "10d3e612-fc06-4163-9744-de41c9a9a307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1363.565701] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Releasing lock "refresh_cache-6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1363.566055] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Instance network_info: |[{"id": "10d3e612-fc06-4163-9744-de41c9a9a307", "address": "fa:16:3e:8b:d3:4e", "network": {"id": "ec789ef1-85c7-4760-b9d7-cffbbdc71557", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-466404986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc80e86b089d4c8c92b721d8fa944fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d177c5b3-a5b1-4c78-854e-7e0dbf341ea1", "external-id": "nsx-vlan-transportzone-54", "segmentation_id": 54, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d3e612-fc", "ovs_interfaceid": "10d3e612-fc06-4163-9744-de41c9a9a307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61986) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1363.566502] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:d3:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd177c5b3-a5b1-4c78-854e-7e0dbf341ea1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10d3e612-fc06-4163-9744-de41c9a9a307', 'vif_model': 'vmxnet3'}] {{(pid=61986) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1363.574062] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Creating folder: Project (bc80e86b089d4c8c92b721d8fa944fb5). Parent ref: group-v252271. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1363.574352] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9175121-a2fa-4211-b24f-3bbe1abb379f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.586483] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Created folder: Project (bc80e86b089d4c8c92b721d8fa944fb5) in parent group-v252271. [ 1363.586661] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Creating folder: Instances. Parent ref: group-v252508. {{(pid=61986) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1363.587166] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0deabdf2-e286-419e-80b5-8d3c56c90c12 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.595493] env[61986]: INFO nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Created folder: Instances in parent group-v252508. [ 1363.595713] env[61986]: DEBUG oslo.service.loopingcall [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1363.595886] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Creating VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1363.596080] env[61986]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84cdefa9-426f-405c-b04a-94622a0887c4 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.614026] env[61986]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1363.614026] env[61986]: value = "task-1160748" [ 1363.614026] env[61986]: _type = "Task" [ 1363.614026] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.621777] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160748, 'name': CreateVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.852919] env[61986]: DEBUG nova.compute.manager [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Received event network-changed-10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1363.853213] env[61986]: DEBUG nova.compute.manager [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Refreshing instance network info cache due to event network-changed-10d3e612-fc06-4163-9744-de41c9a9a307. {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1363.853447] env[61986]: DEBUG oslo_concurrency.lockutils [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] Acquiring lock "refresh_cache-6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1363.853598] env[61986]: DEBUG oslo_concurrency.lockutils [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] Acquired lock "refresh_cache-6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1363.853759] env[61986]: DEBUG nova.network.neutron [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Refreshing network info cache for port 10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1364.123714] env[61986]: DEBUG oslo_vmware.api [-] Task: {'id': task-1160748, 'name': CreateVM_Task, 'duration_secs': 0.31056} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.124095] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Created VM on the ESX host {{(pid=61986) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1364.124601] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1364.124787] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1364.125130] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1364.125380] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5fcafa5-c3cb-4a56-838a-55f7896f2c1c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.129882] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1364.129882] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522d6670-25f3-1d4c-c9f0-332b0320cd2c" [ 1364.129882] env[61986]: _type = "Task" [ 1364.129882] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.138966] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522d6670-25f3-1d4c-c9f0-332b0320cd2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.529157] env[61986]: DEBUG nova.network.neutron [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Updated VIF entry in instance network info cache for port 10d3e612-fc06-4163-9744-de41c9a9a307. {{(pid=61986) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1364.529609] env[61986]: DEBUG nova.network.neutron [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Updating instance_info_cache with network_info: [{"id": "10d3e612-fc06-4163-9744-de41c9a9a307", "address": "fa:16:3e:8b:d3:4e", "network": {"id": "ec789ef1-85c7-4760-b9d7-cffbbdc71557", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-466404986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc80e86b089d4c8c92b721d8fa944fb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d177c5b3-a5b1-4c78-854e-7e0dbf341ea1", "external-id": "nsx-vlan-transportzone-54", "segmentation_id": 54, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d3e612-fc", "ovs_interfaceid": "10d3e612-fc06-4163-9744-de41c9a9a307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1364.641108] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]522d6670-25f3-1d4c-c9f0-332b0320cd2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010585} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.641380] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1364.641623] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Processing image 7b73cf0e-555e-4258-a7ba-6cecbc90b6c5 {{(pid=61986) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1364.641853] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1364.642084] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1364.642322] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1364.642599] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf4af37f-a20d-4ebf-9e80-6ee9837fe9a3 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.650847] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61986) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1364.651054] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61986) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1364.651790] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ec7be5a-ce57-4b5a-b979-c4eaef0fa264 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.656932] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1364.656932] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c6abf3-6a73-3fb4-5e6e-5c19899b11e1" [ 1364.656932] env[61986]: _type = "Task" [ 1364.656932] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.664103] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c6abf3-6a73-3fb4-5e6e-5c19899b11e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.033134] env[61986]: DEBUG oslo_concurrency.lockutils [req-0c881726-3052-46b3-815d-0bde962bd928 req-44d24b8b-62cf-4071-9565-b9e831472446 service nova] Releasing lock "refresh_cache-6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1365.167574] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]52c6abf3-6a73-3fb4-5e6e-5c19899b11e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009659} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.168358] env[61986]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b053ce77-a8b3-425d-a75c-8783edbab9da {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.173165] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1365.173165] env[61986]: value = "session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521689c4-80a3-2b6c-0e0a-865481edf7e6" [ 1365.173165] env[61986]: _type = "Task" [ 1365.173165] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.181112] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521689c4-80a3-2b6c-0e0a-865481edf7e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.683762] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': session[5210eb54-a7a4-b41c-817c-9d8b0b10d56f]521689c4-80a3-2b6c-0e0a-865481edf7e6, 'name': SearchDatastore_Task, 'duration_secs': 0.010784} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.684041] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk" {{(pid=61986) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1365.684373] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc/6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1365.684705] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b5e05b1-a83a-4a72-9d10-5a705936b38c {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.691249] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1365.691249] env[61986]: value = "task-1160749" [ 1365.691249] env[61986]: _type = "Task" [ 1365.691249] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.699018] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160749, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.200812] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160749, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46903} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.201207] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5/7b73cf0e-555e-4258-a7ba-6cecbc90b6c5.vmdk to [datastore2] 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc/6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc.vmdk {{(pid=61986) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1366.201314] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Extending root virtual disk to 1048576 {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1366.201535] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8eb74d88-81ff-4741-a033-e20ea988127b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.207311] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1366.207311] env[61986]: value = "task-1160750" [ 1366.207311] env[61986]: _type = "Task" [ 1366.207311] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.214307] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.717292] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066676} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.717562] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Extended root virtual disk {{(pid=61986) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1366.718334] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d297e332-477f-426a-8a89-38ae41d80b54 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.738971] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc/6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1366.739206] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3700d6a-b883-41e8-af20-3dacaea40915 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.757516] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1366.757516] env[61986]: value = "task-1160751" [ 1366.757516] env[61986]: _type = "Task" [ 1366.757516] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.764541] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160751, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.267522] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160751, 'name': ReconfigVM_Task, 'duration_secs': 0.28891} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.267862] env[61986]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc/6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc.vmdk or device None with type sparse {{(pid=61986) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1367.268404] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c583fd94-9c35-4771-a53b-88999f5ec078 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.274462] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1367.274462] env[61986]: value = "task-1160752" [ 1367.274462] env[61986]: _type = "Task" [ 1367.274462] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.282713] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160752, 'name': Rename_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.785019] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160752, 'name': Rename_Task, 'duration_secs': 0.129201} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.785319] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Powering on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1367.785581] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e23f5e8-e607-4ce8-874e-558867cf8642 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.791666] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1367.791666] env[61986]: value = "task-1160753" [ 1367.791666] env[61986]: _type = "Task" [ 1367.791666] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.798675] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.302227] env[61986]: DEBUG oslo_vmware.api [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160753, 'name': PowerOnVM_Task, 'duration_secs': 0.449633} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.302597] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Powered on the VM {{(pid=61986) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1368.302677] env[61986]: INFO nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Took 6.61 seconds to spawn the instance on the hypervisor. [ 1368.302859] env[61986]: DEBUG nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Checking state {{(pid=61986) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1368.303655] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccd5a87-e98d-4d61-a39c-baec5797eb60 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.821802] env[61986]: INFO nova.compute.manager [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Took 11.28 seconds to build instance. [ 1369.322929] env[61986]: DEBUG oslo_concurrency.lockutils [None req-b8deb5dc-e607-4eda-b127-d97b06c9173e tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.789s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1370.042051] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.042364] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.042593] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.042784] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.042990] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1370.045256] env[61986]: INFO nova.compute.manager [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Terminating instance [ 1370.047079] env[61986]: DEBUG nova.compute.manager [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Start destroying the instance on the hypervisor. {{(pid=61986) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1370.047284] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Destroying instance {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1370.048134] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d37c27-d136-41c9-bcf0-d626e765c37f {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.055738] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Powering off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1370.056056] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32b78573-4b05-48d2-861d-b10498bc7410 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.063840] env[61986]: DEBUG oslo_vmware.api [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1370.063840] env[61986]: value = "task-1160754" [ 1370.063840] env[61986]: _type = "Task" [ 1370.063840] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.072167] env[61986]: DEBUG oslo_vmware.api [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.574185] env[61986]: DEBUG oslo_vmware.api [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160754, 'name': PowerOffVM_Task, 'duration_secs': 0.180792} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.574561] env[61986]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Powered off the VM {{(pid=61986) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1370.574657] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Unregistering the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1370.574905] env[61986]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db8fd0ea-f693-4a39-aa06-9e81ef05ebe7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.638550] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Unregistered the VM {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1370.638766] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Deleting contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1370.638951] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Deleting the datastore file [datastore2] 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1370.639249] env[61986]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70bacad8-971e-415d-91cd-fa3890b6d8de {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.645203] env[61986]: DEBUG oslo_vmware.api [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for the task: (returnval){ [ 1370.645203] env[61986]: value = "task-1160756" [ 1370.645203] env[61986]: _type = "Task" [ 1370.645203] env[61986]: } to complete. {{(pid=61986) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.652794] env[61986]: DEBUG oslo_vmware.api [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.154832] env[61986]: DEBUG oslo_vmware.api [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Task: {'id': task-1160756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154272} completed successfully. {{(pid=61986) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.156630] env[61986]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Deleted the datastore file {{(pid=61986) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1371.156630] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Deleted contents of the VM from datastore datastore2 {{(pid=61986) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1371.156630] env[61986]: DEBUG nova.virt.vmwareapi.vmops [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Instance destroyed {{(pid=61986) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1371.156630] env[61986]: INFO nova.compute.manager [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1371.156630] env[61986]: DEBUG oslo.service.loopingcall [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61986) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1371.156630] env[61986]: DEBUG nova.compute.manager [-] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Deallocating network for instance {{(pid=61986) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1371.156630] env[61986]: DEBUG nova.network.neutron [-] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] deallocate_for_instance() {{(pid=61986) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1371.421319] env[61986]: DEBUG nova.compute.manager [req-37e19de9-174c-40f9-b987-6f8ce3119bc4 req-c874f22c-56d4-4945-8570-7820ec4b6f22 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Received event network-vif-deleted-10d3e612-fc06-4163-9744-de41c9a9a307 {{(pid=61986) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1371.421571] env[61986]: INFO nova.compute.manager [req-37e19de9-174c-40f9-b987-6f8ce3119bc4 req-c874f22c-56d4-4945-8570-7820ec4b6f22 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Neutron deleted interface 10d3e612-fc06-4163-9744-de41c9a9a307; detaching it from the instance and deleting it from the info cache [ 1371.421757] env[61986]: DEBUG nova.network.neutron [req-37e19de9-174c-40f9-b987-6f8ce3119bc4 req-c874f22c-56d4-4945-8570-7820ec4b6f22 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1371.894937] env[61986]: DEBUG nova.network.neutron [-] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Updating instance_info_cache with network_info: [] {{(pid=61986) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1371.924402] env[61986]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d8d33b8b-e05d-4f3f-b3e3-6387e4a4558d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.933831] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e2b428-e232-42f6-ae22-cb12bac1f834 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.955352] env[61986]: DEBUG nova.compute.manager [req-37e19de9-174c-40f9-b987-6f8ce3119bc4 req-c874f22c-56d4-4945-8570-7820ec4b6f22 service nova] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Detach interface failed, port_id=10d3e612-fc06-4163-9744-de41c9a9a307, reason: Instance 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc could not be found. {{(pid=61986) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1372.397937] env[61986]: INFO nova.compute.manager [-] [instance: 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc] Took 1.24 seconds to deallocate network for instance. [ 1372.905996] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.906359] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.906618] env[61986]: DEBUG nova.objects.instance [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lazy-loading 'resources' on Instance uuid 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc {{(pid=61986) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1373.440573] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460de9bc-49fc-4bb2-8d12-546d751a28c9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.447815] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330ff396-a567-45dd-9e19-76579e500c82 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.476547] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cb3c36-2694-4057-a167-494bfb9ac3a9 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.483670] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ad9349-2890-4361-9389-8fb716a98637 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.497397] env[61986]: DEBUG nova.compute.provider_tree [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1374.000277] env[61986]: DEBUG nova.scheduler.client.report [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1374.505902] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.525302] env[61986]: INFO nova.scheduler.client.report [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Deleted allocations for instance 6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc [ 1375.033581] env[61986]: DEBUG oslo_concurrency.lockutils [None req-bb83eb49-4d47-493b-bca0-cabc12dd2386 tempest-ServerMetadataNegativeTestJSON-257189895 tempest-ServerMetadataNegativeTestJSON-257189895-project-member] Lock "6aaa8bbc-1a6d-43c4-872e-feb2648dd0cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.991s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1391.214036] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.214421] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.214421] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61986) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1391.214766] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.717940] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.718169] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1391.718269] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1391.718428] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61986) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1391.719360] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2753019-1b35-4425-b6eb-f00f25b5fc2b {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.727620] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24df99a-9cdd-4a12-bd9b-faf5dd9eeb7e {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.741099] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf68d9a-4d65-4990-88c2-547e7aa754e7 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.747124] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bc10e0-6701-44c6-a42e-ce4991cf7ea8 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.776278] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181085MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=61986) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1391.776475] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.776674] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1392.796297] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1392.796615] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61986) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1392.809332] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef70439-7d69-49a7-9880-292336039afe {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.816910] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7723baa7-6969-4b7a-963f-d1e1922bd297 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.845208] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0cd6b5-c2ce-4b3c-995d-926561e6736d {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.851617] env[61986]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2350fc0c-781d-41d5-b34c-7dbe5a36aa06 {{(pid=61986) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.863800] env[61986]: DEBUG nova.compute.provider_tree [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed in ProviderTree for provider: 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d {{(pid=61986) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1393.366727] env[61986]: DEBUG nova.scheduler.client.report [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Inventory has not changed for provider 2abb0eb7-c8d4-4d3b-8ee3-70b19852b51d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61986) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1393.871861] env[61986]: DEBUG nova.compute.resource_tracker [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61986) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1393.872318] env[61986]: DEBUG oslo_concurrency.lockutils [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.095s {{(pid=61986) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.872340] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.872690] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Starting heal instance info cache {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1394.872690] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Rebuilding the list of instances to heal {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1395.375627] env[61986]: DEBUG nova.compute.manager [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Didn't find any instances for network info cache update. {{(pid=61986) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1395.375962] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1395.376095] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1395.713756] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1395.713987] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1396.219309] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.214629] env[61986]: DEBUG oslo_service.periodic_task [None req-5d43bee9-3b18-4c2b-8272-fb9c5b77c06e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61986) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}